var/home/core/zuul-output/0000755000175000017500000000000015066706672014544 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015066724077015507 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005247155515066724071017723 0ustar rootrootSep 30 08:02:55 crc systemd[1]: Starting Kubernetes Kubelet... Sep 30 08:02:55 crc restorecon[4666]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:55 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:56 crc restorecon[4666]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 08:02:56 crc restorecon[4666]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Sep 30 08:02:57 crc kubenswrapper[4810]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 08:02:57 crc kubenswrapper[4810]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Sep 30 08:02:57 crc kubenswrapper[4810]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 08:02:57 crc kubenswrapper[4810]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 08:02:57 crc kubenswrapper[4810]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 30 08:02:57 crc kubenswrapper[4810]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.025893 4810 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031733 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031788 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031799 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031808 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031819 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031829 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031839 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031851 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031861 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031871 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031881 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031890 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031900 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031909 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031919 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031929 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031936 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031944 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031954 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031963 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031973 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031983 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.031993 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032003 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032012 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032022 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032030 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032037 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032045 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032053 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032061 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032068 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032075 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032097 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032107 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032116 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032123 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032131 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032139 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032147 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032157 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032165 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032173 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032180 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032189 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032196 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032204 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032212 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032220 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032232 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032246 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032257 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032293 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032302 4810 feature_gate.go:330] unrecognized feature gate: Example Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032312 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032320 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032333 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032343 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032352 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032360 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032368 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032376 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032385 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032393 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032400 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032410 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032418 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032427 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032435 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032442 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.032451 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.033961 4810 flags.go:64] FLAG: --address="0.0.0.0" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.033987 4810 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034006 4810 flags.go:64] FLAG: --anonymous-auth="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034025 4810 flags.go:64] FLAG: --application-metrics-count-limit="100" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034039 4810 flags.go:64] FLAG: --authentication-token-webhook="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034048 4810 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034061 4810 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034072 4810 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034082 4810 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034092 4810 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034102 4810 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034117 4810 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034127 4810 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034136 4810 flags.go:64] FLAG: --cgroup-root="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034145 4810 flags.go:64] FLAG: --cgroups-per-qos="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034154 4810 flags.go:64] FLAG: --client-ca-file="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034162 4810 flags.go:64] FLAG: --cloud-config="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034171 4810 flags.go:64] FLAG: --cloud-provider="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034180 4810 flags.go:64] FLAG: --cluster-dns="[]" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034192 4810 flags.go:64] FLAG: --cluster-domain="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034201 4810 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034210 4810 flags.go:64] FLAG: --config-dir="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034219 4810 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034229 4810 flags.go:64] FLAG: --container-log-max-files="5" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034242 4810 flags.go:64] FLAG: --container-log-max-size="10Mi" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034251 4810 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034261 4810 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034296 4810 flags.go:64] FLAG: --containerd-namespace="k8s.io" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034306 4810 flags.go:64] FLAG: --contention-profiling="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034315 4810 flags.go:64] FLAG: --cpu-cfs-quota="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034324 4810 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034334 4810 flags.go:64] FLAG: --cpu-manager-policy="none" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034343 4810 flags.go:64] FLAG: --cpu-manager-policy-options="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034355 4810 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034364 4810 flags.go:64] FLAG: --enable-controller-attach-detach="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034373 4810 flags.go:64] FLAG: --enable-debugging-handlers="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034382 4810 flags.go:64] FLAG: --enable-load-reader="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034391 4810 flags.go:64] FLAG: --enable-server="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034400 4810 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034413 4810 flags.go:64] FLAG: --event-burst="100" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034422 4810 flags.go:64] FLAG: --event-qps="50" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034431 4810 flags.go:64] FLAG: --event-storage-age-limit="default=0" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034441 4810 flags.go:64] FLAG: --event-storage-event-limit="default=0" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034450 4810 flags.go:64] FLAG: --eviction-hard="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034461 4810 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034470 4810 flags.go:64] FLAG: --eviction-minimum-reclaim="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034479 4810 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034491 4810 flags.go:64] FLAG: --eviction-soft="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034501 4810 flags.go:64] FLAG: --eviction-soft-grace-period="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034509 4810 flags.go:64] FLAG: --exit-on-lock-contention="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034518 4810 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034527 4810 flags.go:64] FLAG: --experimental-mounter-path="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034536 4810 flags.go:64] FLAG: --fail-cgroupv1="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034545 4810 flags.go:64] FLAG: --fail-swap-on="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034554 4810 flags.go:64] FLAG: --feature-gates="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034565 4810 flags.go:64] FLAG: --file-check-frequency="20s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034574 4810 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034584 4810 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034594 4810 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034604 4810 flags.go:64] FLAG: --healthz-port="10248" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034614 4810 flags.go:64] FLAG: --help="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034623 4810 flags.go:64] FLAG: --hostname-override="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034632 4810 flags.go:64] FLAG: --housekeeping-interval="10s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034641 4810 flags.go:64] FLAG: --http-check-frequency="20s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034651 4810 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034659 4810 flags.go:64] FLAG: --image-credential-provider-config="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034668 4810 flags.go:64] FLAG: --image-gc-high-threshold="85" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034677 4810 flags.go:64] FLAG: --image-gc-low-threshold="80" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034686 4810 flags.go:64] FLAG: --image-service-endpoint="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034695 4810 flags.go:64] FLAG: --kernel-memcg-notification="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034704 4810 flags.go:64] FLAG: --kube-api-burst="100" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034713 4810 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034723 4810 flags.go:64] FLAG: --kube-api-qps="50" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034732 4810 flags.go:64] FLAG: --kube-reserved="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034742 4810 flags.go:64] FLAG: --kube-reserved-cgroup="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034750 4810 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034760 4810 flags.go:64] FLAG: --kubelet-cgroups="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034769 4810 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034778 4810 flags.go:64] FLAG: --lock-file="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034787 4810 flags.go:64] FLAG: --log-cadvisor-usage="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034795 4810 flags.go:64] FLAG: --log-flush-frequency="5s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034805 4810 flags.go:64] FLAG: --log-json-info-buffer-size="0" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034819 4810 flags.go:64] FLAG: --log-json-split-stream="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034829 4810 flags.go:64] FLAG: --log-text-info-buffer-size="0" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034838 4810 flags.go:64] FLAG: --log-text-split-stream="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034848 4810 flags.go:64] FLAG: --logging-format="text" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034857 4810 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034867 4810 flags.go:64] FLAG: --make-iptables-util-chains="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034876 4810 flags.go:64] FLAG: --manifest-url="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034885 4810 flags.go:64] FLAG: --manifest-url-header="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034899 4810 flags.go:64] FLAG: --max-housekeeping-interval="15s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034908 4810 flags.go:64] FLAG: --max-open-files="1000000" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034920 4810 flags.go:64] FLAG: --max-pods="110" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034929 4810 flags.go:64] FLAG: --maximum-dead-containers="-1" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034938 4810 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034947 4810 flags.go:64] FLAG: --memory-manager-policy="None" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034956 4810 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034965 4810 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034989 4810 flags.go:64] FLAG: --node-ip="192.168.126.11" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.034999 4810 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035026 4810 flags.go:64] FLAG: --node-status-max-images="50" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035035 4810 flags.go:64] FLAG: --node-status-update-frequency="10s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035044 4810 flags.go:64] FLAG: --oom-score-adj="-999" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035053 4810 flags.go:64] FLAG: --pod-cidr="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035062 4810 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035077 4810 flags.go:64] FLAG: --pod-manifest-path="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035089 4810 flags.go:64] FLAG: --pod-max-pids="-1" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035129 4810 flags.go:64] FLAG: --pods-per-core="0" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035139 4810 flags.go:64] FLAG: --port="10250" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035149 4810 flags.go:64] FLAG: --protect-kernel-defaults="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035158 4810 flags.go:64] FLAG: --provider-id="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035167 4810 flags.go:64] FLAG: --qos-reserved="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035176 4810 flags.go:64] FLAG: --read-only-port="10255" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035185 4810 flags.go:64] FLAG: --register-node="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035194 4810 flags.go:64] FLAG: --register-schedulable="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035203 4810 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035220 4810 flags.go:64] FLAG: --registry-burst="10" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035229 4810 flags.go:64] FLAG: --registry-qps="5" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035238 4810 flags.go:64] FLAG: --reserved-cpus="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035248 4810 flags.go:64] FLAG: --reserved-memory="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035285 4810 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035295 4810 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035305 4810 flags.go:64] FLAG: --rotate-certificates="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035314 4810 flags.go:64] FLAG: --rotate-server-certificates="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035322 4810 flags.go:64] FLAG: --runonce="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035331 4810 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035341 4810 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035350 4810 flags.go:64] FLAG: --seccomp-default="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035359 4810 flags.go:64] FLAG: --serialize-image-pulls="true" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035368 4810 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035381 4810 flags.go:64] FLAG: --storage-driver-db="cadvisor" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035391 4810 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035400 4810 flags.go:64] FLAG: --storage-driver-password="root" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035409 4810 flags.go:64] FLAG: --storage-driver-secure="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035418 4810 flags.go:64] FLAG: --storage-driver-table="stats" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035427 4810 flags.go:64] FLAG: --storage-driver-user="root" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035437 4810 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035446 4810 flags.go:64] FLAG: --sync-frequency="1m0s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035456 4810 flags.go:64] FLAG: --system-cgroups="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035465 4810 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035480 4810 flags.go:64] FLAG: --system-reserved-cgroup="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035489 4810 flags.go:64] FLAG: --tls-cert-file="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035498 4810 flags.go:64] FLAG: --tls-cipher-suites="[]" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035511 4810 flags.go:64] FLAG: --tls-min-version="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035520 4810 flags.go:64] FLAG: --tls-private-key-file="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035529 4810 flags.go:64] FLAG: --topology-manager-policy="none" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035538 4810 flags.go:64] FLAG: --topology-manager-policy-options="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035548 4810 flags.go:64] FLAG: --topology-manager-scope="container" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035557 4810 flags.go:64] FLAG: --v="2" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035570 4810 flags.go:64] FLAG: --version="false" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035583 4810 flags.go:64] FLAG: --vmodule="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035594 4810 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.035604 4810 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035820 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035830 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035839 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035847 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035856 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035864 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035872 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035879 4810 feature_gate.go:330] unrecognized feature gate: Example Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035887 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035899 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035906 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035914 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035922 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035929 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035937 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035945 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035953 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035961 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035969 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035977 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035985 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.035993 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036001 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036008 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036016 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036024 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036031 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036039 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036047 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036055 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036063 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036070 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036078 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036086 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036093 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036101 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036109 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036116 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036125 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036133 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036141 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036151 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036159 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036169 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036179 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036187 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036195 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036203 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036211 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036219 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036226 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036235 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036243 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036250 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036258 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036292 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036301 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036309 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036316 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036327 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036336 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036344 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036352 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036360 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036370 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036380 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036390 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036399 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036407 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036415 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.036423 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.037485 4810 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.052212 4810 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.052327 4810 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052471 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052496 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052507 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052518 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052527 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052536 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052546 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052555 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052567 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052583 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052593 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052604 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052615 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052624 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052633 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052644 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052660 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052672 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052684 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052698 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052710 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052721 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052733 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052744 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052755 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052763 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052772 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052782 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052791 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052801 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052810 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052820 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052829 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052838 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052849 4810 feature_gate.go:330] unrecognized feature gate: Example Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052858 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052866 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052875 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052883 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052892 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052901 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052909 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052917 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052926 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052937 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052950 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052959 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052969 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052978 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052987 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.052996 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053006 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053015 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053024 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053034 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053043 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053053 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053065 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053076 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053088 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053100 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053110 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053121 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053135 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053148 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053159 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053169 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053180 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053191 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053202 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053215 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.053232 4810 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053661 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053679 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053690 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053700 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053709 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053718 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053727 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053736 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053746 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053755 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053764 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053772 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053784 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053795 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053804 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053813 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053822 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053831 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053840 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053848 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053857 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053866 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053874 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053883 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053892 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053903 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053915 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053924 4810 feature_gate.go:330] unrecognized feature gate: Example Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053933 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053942 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053951 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053959 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053968 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053977 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053989 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.053998 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054007 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054016 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054025 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054034 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054043 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054051 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054060 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054068 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054080 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054091 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054101 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054110 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054121 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054131 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054140 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054150 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054158 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054167 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054176 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054184 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054193 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054203 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054211 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054220 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054229 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054237 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054246 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054257 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054299 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054310 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054319 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054329 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054338 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054347 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.054371 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.054385 4810 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.054749 4810 server.go:940] "Client rotation is on, will bootstrap in background" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.062344 4810 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.062515 4810 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.064355 4810 server.go:997] "Starting client certificate rotation" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.064409 4810 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.064769 4810 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-29 10:09:52.123907787 +0000 UTC Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.064894 4810 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1442h6m55.059022371s for next certificate rotation Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.093370 4810 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.098213 4810 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.120744 4810 log.go:25] "Validated CRI v1 runtime API" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.161500 4810 log.go:25] "Validated CRI v1 image API" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.163803 4810 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.172494 4810 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-09-30-07-41-45-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.172557 4810 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:49 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.204756 4810 manager.go:217] Machine: {Timestamp:2025-09-30 08:02:57.199703241 +0000 UTC m=+0.651902538 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:3de9b6fb-1269-4c3c-a6c1-7b041590cd04 BootID:441a0caf-7730-4ad7-a34e-ac02f226b5b0 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:49 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:04:bd:33 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:04:bd:33 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:5c:db:26 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:c9:56:35 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:45:e4:05 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:84:a4:ca Speed:-1 Mtu:1496} {Name:eth10 MacAddress:e2:da:a9:80:a5:90 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:22:38:54:82:69:f8 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.205169 4810 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.205498 4810 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.206057 4810 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.206408 4810 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.206469 4810 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.206814 4810 topology_manager.go:138] "Creating topology manager with none policy" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.206834 4810 container_manager_linux.go:303] "Creating device plugin manager" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.207434 4810 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.207486 4810 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.207725 4810 state_mem.go:36] "Initialized new in-memory state store" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.207869 4810 server.go:1245] "Using root directory" path="/var/lib/kubelet" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.210893 4810 kubelet.go:418] "Attempting to sync node with API server" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.210927 4810 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.210965 4810 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.210986 4810 kubelet.go:324] "Adding apiserver pod source" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.211005 4810 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.216359 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.216476 4810 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.216841 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.203:6443: connect: connection refused" logger="UnhandledError" Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.216359 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.217083 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.203:6443: connect: connection refused" logger="UnhandledError" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.218011 4810 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.219441 4810 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221573 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221615 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221629 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221643 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221665 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221678 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221692 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221713 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221731 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221745 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221767 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.221785 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.223016 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.223886 4810 server.go:1280] "Started kubelet" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.225884 4810 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 30 08:02:57 crc systemd[1]: Started Kubernetes Kubelet. Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.226216 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.226145 4810 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.226960 4810 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.227819 4810 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.227876 4810 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.227941 4810 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 11:59:26.780774231 +0000 UTC Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.227997 4810 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1155h56m29.552781762s for next certificate rotation Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.228191 4810 volume_manager.go:287] "The desired_state_of_world populator starts" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.228215 4810 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.228418 4810 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.228407 4810 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.232070 4810 factory.go:55] Registering systemd factory Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.232114 4810 factory.go:221] Registration of the systemd container factory successfully Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.235858 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.236262 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.203:6443: connect: connection refused" logger="UnhandledError" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.236365 4810 factory.go:153] Registering CRI-O factory Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.236410 4810 factory.go:221] Registration of the crio container factory successfully Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.236527 4810 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.236565 4810 factory.go:103] Registering Raw factory Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.236593 4810 manager.go:1196] Started watching for new ooms in manager Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.237931 4810 manager.go:319] Starting recovery of all containers Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.239817 4810 server.go:460] "Adding debug handlers to kubelet server" Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.241106 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.203:6443: connect: connection refused" interval="200ms" Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.252036 4810 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.203:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a00b3ed690e51 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-09-30 08:02:57.223831121 +0000 UTC m=+0.676030418,LastTimestamp:2025-09-30 08:02:57.223831121 +0000 UTC m=+0.676030418,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261289 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261371 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261395 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261415 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261435 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261454 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261475 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261495 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261537 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261560 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261580 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261599 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261617 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261642 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261661 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261681 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261704 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261733 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261763 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261781 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261800 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261820 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261838 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261889 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261916 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261937 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261963 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.261985 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262007 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262028 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262047 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262068 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262092 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262112 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262134 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262156 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262176 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262199 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262220 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262241 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262261 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262305 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262328 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262348 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262368 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262389 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262408 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262427 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262450 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262471 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262493 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262512 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262538 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262558 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262580 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262602 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262623 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262645 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262664 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262682 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262701 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262720 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262740 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262758 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262781 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262800 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262819 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262841 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262860 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262880 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262899 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262922 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262940 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262960 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262979 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.262997 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263016 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263035 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263054 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263076 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263096 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263115 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263136 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263156 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263175 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263195 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263214 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263234 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263253 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263295 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263317 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263338 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263357 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263377 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263399 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263420 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263444 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263466 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263486 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263506 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263525 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263545 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263564 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263584 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263613 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263636 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263657 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263678 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263704 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263726 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263747 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263767 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263789 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263809 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263831 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.263852 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267318 4810 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267364 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267387 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267410 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267430 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267451 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267470 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267490 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267512 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267533 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267553 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267576 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267599 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267618 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267638 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267658 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267679 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267698 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267719 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267740 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267761 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267781 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267853 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267876 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267897 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267917 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267939 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267958 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267979 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.267999 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268019 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268039 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268061 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268135 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268158 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268178 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268198 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268240 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268260 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268305 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268324 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268343 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268362 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268382 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268446 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268472 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268495 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268534 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268553 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268575 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268595 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268615 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268649 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268669 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268689 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268707 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268726 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268745 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268765 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268785 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268804 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268825 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268844 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268864 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268884 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268902 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268923 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268942 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268961 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.268982 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269002 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269031 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269064 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269083 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269101 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269121 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269147 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269165 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269185 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269204 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269223 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269242 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269286 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269308 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269326 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269350 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269370 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269389 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269408 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269426 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269446 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269465 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269484 4810 reconstruct.go:97] "Volume reconstruction finished" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.269498 4810 reconciler.go:26] "Reconciler: start to sync state" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.275836 4810 manager.go:324] Recovery completed Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.288588 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.290798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.290872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.290893 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.291752 4810 cpu_manager.go:225] "Starting CPU manager" policy="none" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.291774 4810 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.291825 4810 state_mem.go:36] "Initialized new in-memory state store" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.301442 4810 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.305044 4810 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.305118 4810 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.305157 4810 kubelet.go:2335] "Starting kubelet main sync loop" Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.305231 4810 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.306207 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.306363 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.203:6443: connect: connection refused" logger="UnhandledError" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.310326 4810 policy_none.go:49] "None policy: Start" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.311505 4810 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.311539 4810 state_mem.go:35] "Initializing new in-memory state store" Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.329386 4810 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.376242 4810 manager.go:334] "Starting Device Plugin manager" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.376383 4810 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.376477 4810 server.go:79] "Starting device plugin registration server" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.377938 4810 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.378002 4810 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.378209 4810 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.378405 4810 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.378420 4810 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.393930 4810 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.406145 4810 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.406301 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.407616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.407678 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.407702 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.407996 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.409190 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.409236 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.409586 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.409661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.409724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.409997 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.410062 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.410098 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.410197 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.410257 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.410306 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.411190 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.411250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.411255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.411282 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.411318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.411329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.411430 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.411739 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.411822 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.412200 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.412264 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.412337 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.412593 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.412755 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.412797 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.413192 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.413250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.413302 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.413744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.413791 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.413816 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.413892 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.413909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.413920 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.414035 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.414055 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.414938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.414989 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.415014 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.442020 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.203:6443: connect: connection refused" interval="400ms" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473126 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473174 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473203 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473226 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473250 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473287 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473307 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473439 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473510 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473553 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473587 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473622 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473653 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473725 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.473784 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.478430 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.479993 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.480044 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.480058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.480101 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.480635 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.203:6443: connect: connection refused" node="crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575133 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575229 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575301 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575335 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575370 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575404 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575409 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575442 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575503 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575516 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575459 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575450 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575428 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575520 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575551 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575608 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575588 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575656 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575691 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575719 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575731 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575765 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575796 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575828 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575836 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575867 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575910 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.575960 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.576009 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.576063 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.681302 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.682813 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.682908 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.682926 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.682968 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.683700 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.203:6443: connect: connection refused" node="crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.752710 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.779706 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.795080 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-7136bdf439c164d1535703d4bed939a49b81b2506a896a9c513119fdd289238e WatchSource:0}: Error finding container 7136bdf439c164d1535703d4bed939a49b81b2506a896a9c513119fdd289238e: Status 404 returned error can't find the container with id 7136bdf439c164d1535703d4bed939a49b81b2506a896a9c513119fdd289238e Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.808915 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-9a89ce7bc27f9818fd3be365ffaf61b03d1aaa9da3fdc24d529905ac6c8eb8be WatchSource:0}: Error finding container 9a89ce7bc27f9818fd3be365ffaf61b03d1aaa9da3fdc24d529905ac6c8eb8be: Status 404 returned error can't find the container with id 9a89ce7bc27f9818fd3be365ffaf61b03d1aaa9da3fdc24d529905ac6c8eb8be Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.816863 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.831419 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: I0930 08:02:57.840849 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 08:02:57 crc kubenswrapper[4810]: E0930 08:02:57.843151 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.203:6443: connect: connection refused" interval="800ms" Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.844604 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ee6f5de65a754bd9f729fd45cc013e701658fc7c972c27b2444e0ed5ea36465e WatchSource:0}: Error finding container ee6f5de65a754bd9f729fd45cc013e701658fc7c972c27b2444e0ed5ea36465e: Status 404 returned error can't find the container with id ee6f5de65a754bd9f729fd45cc013e701658fc7c972c27b2444e0ed5ea36465e Sep 30 08:02:57 crc kubenswrapper[4810]: W0930 08:02:57.860713 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-e6c62dde50ac2b7c60e3aa5782debc0cf097e5ce4f6a998373992fd390f1a941 WatchSource:0}: Error finding container e6c62dde50ac2b7c60e3aa5782debc0cf097e5ce4f6a998373992fd390f1a941: Status 404 returned error can't find the container with id e6c62dde50ac2b7c60e3aa5782debc0cf097e5ce4f6a998373992fd390f1a941 Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.084682 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.086315 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.086360 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.086374 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.086403 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 08:02:58 crc kubenswrapper[4810]: E0930 08:02:58.086710 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.203:6443: connect: connection refused" node="crc" Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.227160 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.310693 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e6c62dde50ac2b7c60e3aa5782debc0cf097e5ce4f6a998373992fd390f1a941"} Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.311815 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ee6f5de65a754bd9f729fd45cc013e701658fc7c972c27b2444e0ed5ea36465e"} Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.312824 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e0f29866abff1f2b837b5e899169b402dee2223be1be6c37f0b67d4788e3907a"} Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.313731 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9a89ce7bc27f9818fd3be365ffaf61b03d1aaa9da3fdc24d529905ac6c8eb8be"} Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.316059 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7136bdf439c164d1535703d4bed939a49b81b2506a896a9c513119fdd289238e"} Sep 30 08:02:58 crc kubenswrapper[4810]: W0930 08:02:58.574231 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:02:58 crc kubenswrapper[4810]: E0930 08:02:58.574382 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.203:6443: connect: connection refused" logger="UnhandledError" Sep 30 08:02:58 crc kubenswrapper[4810]: E0930 08:02:58.644805 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.203:6443: connect: connection refused" interval="1.6s" Sep 30 08:02:58 crc kubenswrapper[4810]: W0930 08:02:58.708539 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:02:58 crc kubenswrapper[4810]: E0930 08:02:58.708646 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.203:6443: connect: connection refused" logger="UnhandledError" Sep 30 08:02:58 crc kubenswrapper[4810]: W0930 08:02:58.794713 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:02:58 crc kubenswrapper[4810]: E0930 08:02:58.794832 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.203:6443: connect: connection refused" logger="UnhandledError" Sep 30 08:02:58 crc kubenswrapper[4810]: W0930 08:02:58.816473 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:02:58 crc kubenswrapper[4810]: E0930 08:02:58.816588 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.203:6443: connect: connection refused" logger="UnhandledError" Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.887689 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.889205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.889242 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.889256 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:58 crc kubenswrapper[4810]: I0930 08:02:58.889527 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 08:02:58 crc kubenswrapper[4810]: E0930 08:02:58.889989 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.203:6443: connect: connection refused" node="crc" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.228042 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.322210 4810 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3" exitCode=0 Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.322301 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3"} Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.322418 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.324113 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.324168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.324189 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.331414 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3"} Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.331465 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.331494 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98"} Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.331520 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69"} Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.331540 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364"} Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.332867 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.332931 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.332952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.334786 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091" exitCode=0 Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.334870 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091"} Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.334943 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.336342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.336373 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.336385 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.338760 4810 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7" exitCode=0 Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.338793 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7"} Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.338909 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.340384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.340415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.340429 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.340810 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.341226 4810 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6277a168b87f31c339a2ad4320707caf113ccf67747f1c2cb6d6ae9f80adba87" exitCode=0 Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.341286 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6277a168b87f31c339a2ad4320707caf113ccf67747f1c2cb6d6ae9f80adba87"} Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.341368 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.342440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.342462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.342472 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.342803 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.342835 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:02:59 crc kubenswrapper[4810]: I0930 08:02:59.342853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.226934 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:03:00 crc kubenswrapper[4810]: E0930 08:03:00.245183 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.203:6443: connect: connection refused" interval="3.2s" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.347780 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b91b227ecf3dbfe23847ac4649941e0d424a4e8e63368b1b5a2c2598d9ac1894"} Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.347849 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6a46e68eef88de27f5f93c67b1a8545fba574725b5bd10afd09688ef453dd66a"} Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.347872 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9075278876c411f05f90291e052877301a61910dc63fecad4b45c1991262da79"} Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.348011 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.349422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.349470 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.349489 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.353667 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc"} Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.353708 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032"} Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.353722 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191"} Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.355886 4810 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0" exitCode=0 Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.355988 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0"} Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.356053 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.357198 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.357220 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.357231 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.359951 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2e2be49cca3b70fcc2cbf49d0e0ee45f96aa0a5939687a96154e60936fd9ded6"} Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.359990 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.360005 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.361097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.361134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.361146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.362100 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.362150 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.362168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:00 crc kubenswrapper[4810]: W0930 08:03:00.477757 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:03:00 crc kubenswrapper[4810]: E0930 08:03:00.477901 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.203:6443: connect: connection refused" logger="UnhandledError" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.490435 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.491778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.491820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.491833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.491862 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 08:03:00 crc kubenswrapper[4810]: E0930 08:03:00.492235 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.203:6443: connect: connection refused" node="crc" Sep 30 08:03:00 crc kubenswrapper[4810]: I0930 08:03:00.599697 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 08:03:00 crc kubenswrapper[4810]: W0930 08:03:00.993707 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:03:00 crc kubenswrapper[4810]: E0930 08:03:00.993802 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.203:6443: connect: connection refused" logger="UnhandledError" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.226898 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:03:01 crc kubenswrapper[4810]: W0930 08:03:01.244871 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.203:6443: connect: connection refused Sep 30 08:03:01 crc kubenswrapper[4810]: E0930 08:03:01.244968 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.203:6443: connect: connection refused" logger="UnhandledError" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.365061 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"46a41799772c60c5ee9148126b752731962fb31857d0ae7630c8a63d76b4680e"} Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.365099 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d"} Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.365136 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.368376 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.368414 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.368427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.369484 4810 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868" exitCode=0 Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.369548 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868"} Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.369559 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.369565 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.369683 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.370352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.370377 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.370387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.370654 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.370685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.370697 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.370779 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.370806 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.370817 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.455494 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.836732 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.837021 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.838655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.838711 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.838733 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:01 crc kubenswrapper[4810]: I0930 08:03:01.896775 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.120433 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.380808 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.380883 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.381063 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f"} Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.381157 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183"} Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.381185 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.381232 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.381192 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce"} Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.382388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.382433 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.382449 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.382757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.382804 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.382826 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.383873 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.383911 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:02 crc kubenswrapper[4810]: I0930 08:03:02.383927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.391110 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad"} Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.391173 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3"} Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.391194 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.391390 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.391394 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.392182 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.393494 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.393539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.393571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.393593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.393634 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.393656 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.393680 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.393719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.393740 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.693225 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.695124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.695183 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.695206 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:03 crc kubenswrapper[4810]: I0930 08:03:03.695248 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 08:03:04 crc kubenswrapper[4810]: I0930 08:03:04.394929 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:04 crc kubenswrapper[4810]: I0930 08:03:04.397987 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:04 crc kubenswrapper[4810]: I0930 08:03:04.398051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:04 crc kubenswrapper[4810]: I0930 08:03:04.398063 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:05 crc kubenswrapper[4810]: I0930 08:03:05.121005 4810 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 30 08:03:05 crc kubenswrapper[4810]: I0930 08:03:05.121125 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 30 08:03:05 crc kubenswrapper[4810]: I0930 08:03:05.772821 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:03:05 crc kubenswrapper[4810]: I0930 08:03:05.773049 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:03:05 crc kubenswrapper[4810]: I0930 08:03:05.773108 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:05 crc kubenswrapper[4810]: I0930 08:03:05.775208 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:05 crc kubenswrapper[4810]: I0930 08:03:05.775316 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:05 crc kubenswrapper[4810]: I0930 08:03:05.775343 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:06 crc kubenswrapper[4810]: I0930 08:03:06.517489 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Sep 30 08:03:06 crc kubenswrapper[4810]: I0930 08:03:06.517823 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:06 crc kubenswrapper[4810]: I0930 08:03:06.519669 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:06 crc kubenswrapper[4810]: I0930 08:03:06.519727 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:06 crc kubenswrapper[4810]: I0930 08:03:06.519750 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:06 crc kubenswrapper[4810]: I0930 08:03:06.797451 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:03:06 crc kubenswrapper[4810]: I0930 08:03:06.797705 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:06 crc kubenswrapper[4810]: I0930 08:03:06.799160 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:06 crc kubenswrapper[4810]: I0930 08:03:06.799230 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:06 crc kubenswrapper[4810]: I0930 08:03:06.799254 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:07 crc kubenswrapper[4810]: E0930 08:03:07.394307 4810 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 30 08:03:07 crc kubenswrapper[4810]: I0930 08:03:07.973338 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:03:07 crc kubenswrapper[4810]: I0930 08:03:07.973614 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:07 crc kubenswrapper[4810]: I0930 08:03:07.975638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:07 crc kubenswrapper[4810]: I0930 08:03:07.975697 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:07 crc kubenswrapper[4810]: I0930 08:03:07.975719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:07 crc kubenswrapper[4810]: I0930 08:03:07.982484 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:03:08 crc kubenswrapper[4810]: I0930 08:03:08.406094 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:08 crc kubenswrapper[4810]: I0930 08:03:08.407807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:08 crc kubenswrapper[4810]: I0930 08:03:08.408010 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:08 crc kubenswrapper[4810]: I0930 08:03:08.408138 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:08 crc kubenswrapper[4810]: I0930 08:03:08.411288 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:03:09 crc kubenswrapper[4810]: I0930 08:03:09.410722 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:09 crc kubenswrapper[4810]: I0930 08:03:09.412105 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:09 crc kubenswrapper[4810]: I0930 08:03:09.412316 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:09 crc kubenswrapper[4810]: I0930 08:03:09.412467 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:10 crc kubenswrapper[4810]: I0930 08:03:10.130733 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Sep 30 08:03:10 crc kubenswrapper[4810]: I0930 08:03:10.131033 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:10 crc kubenswrapper[4810]: I0930 08:03:10.132773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:10 crc kubenswrapper[4810]: I0930 08:03:10.132864 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:10 crc kubenswrapper[4810]: I0930 08:03:10.132884 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:11 crc kubenswrapper[4810]: W0930 08:03:11.497614 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Sep 30 08:03:11 crc kubenswrapper[4810]: I0930 08:03:11.497748 4810 trace.go:236] Trace[358438023]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 08:03:01.495) (total time: 10002ms): Sep 30 08:03:11 crc kubenswrapper[4810]: Trace[358438023]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:03:11.497) Sep 30 08:03:11 crc kubenswrapper[4810]: Trace[358438023]: [10.00200891s] [10.00200891s] END Sep 30 08:03:11 crc kubenswrapper[4810]: E0930 08:03:11.497784 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Sep 30 08:03:11 crc kubenswrapper[4810]: I0930 08:03:11.895484 4810 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Sep 30 08:03:11 crc kubenswrapper[4810]: I0930 08:03:11.895593 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Sep 30 08:03:11 crc kubenswrapper[4810]: I0930 08:03:11.902152 4810 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Sep 30 08:03:11 crc kubenswrapper[4810]: I0930 08:03:11.902227 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Sep 30 08:03:12 crc kubenswrapper[4810]: I0930 08:03:12.420817 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 30 08:03:12 crc kubenswrapper[4810]: I0930 08:03:12.423800 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="46a41799772c60c5ee9148126b752731962fb31857d0ae7630c8a63d76b4680e" exitCode=255 Sep 30 08:03:12 crc kubenswrapper[4810]: I0930 08:03:12.423900 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"46a41799772c60c5ee9148126b752731962fb31857d0ae7630c8a63d76b4680e"} Sep 30 08:03:12 crc kubenswrapper[4810]: I0930 08:03:12.424144 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:12 crc kubenswrapper[4810]: I0930 08:03:12.425339 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:12 crc kubenswrapper[4810]: I0930 08:03:12.425416 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:12 crc kubenswrapper[4810]: I0930 08:03:12.425443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:12 crc kubenswrapper[4810]: I0930 08:03:12.426381 4810 scope.go:117] "RemoveContainer" containerID="46a41799772c60c5ee9148126b752731962fb31857d0ae7630c8a63d76b4680e" Sep 30 08:03:13 crc kubenswrapper[4810]: I0930 08:03:13.430226 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 30 08:03:13 crc kubenswrapper[4810]: I0930 08:03:13.433541 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54"} Sep 30 08:03:13 crc kubenswrapper[4810]: I0930 08:03:13.433769 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:13 crc kubenswrapper[4810]: I0930 08:03:13.435094 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:13 crc kubenswrapper[4810]: I0930 08:03:13.435137 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:13 crc kubenswrapper[4810]: I0930 08:03:13.435155 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:14 crc kubenswrapper[4810]: I0930 08:03:14.438953 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Sep 30 08:03:14 crc kubenswrapper[4810]: I0930 08:03:14.439757 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 30 08:03:14 crc kubenswrapper[4810]: I0930 08:03:14.442230 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54" exitCode=255 Sep 30 08:03:14 crc kubenswrapper[4810]: I0930 08:03:14.442315 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54"} Sep 30 08:03:14 crc kubenswrapper[4810]: I0930 08:03:14.442442 4810 scope.go:117] "RemoveContainer" containerID="46a41799772c60c5ee9148126b752731962fb31857d0ae7630c8a63d76b4680e" Sep 30 08:03:14 crc kubenswrapper[4810]: I0930 08:03:14.442665 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:14 crc kubenswrapper[4810]: I0930 08:03:14.444079 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:14 crc kubenswrapper[4810]: I0930 08:03:14.444124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:14 crc kubenswrapper[4810]: I0930 08:03:14.444144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:14 crc kubenswrapper[4810]: I0930 08:03:14.445095 4810 scope.go:117] "RemoveContainer" containerID="a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54" Sep 30 08:03:14 crc kubenswrapper[4810]: E0930 08:03:14.445455 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 08:03:14 crc kubenswrapper[4810]: I0930 08:03:14.635429 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:03:15 crc kubenswrapper[4810]: I0930 08:03:15.121539 4810 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 30 08:03:15 crc kubenswrapper[4810]: I0930 08:03:15.121651 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 30 08:03:15 crc kubenswrapper[4810]: I0930 08:03:15.447861 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Sep 30 08:03:15 crc kubenswrapper[4810]: I0930 08:03:15.450533 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:15 crc kubenswrapper[4810]: I0930 08:03:15.451987 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:15 crc kubenswrapper[4810]: I0930 08:03:15.452049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:15 crc kubenswrapper[4810]: I0930 08:03:15.452072 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:15 crc kubenswrapper[4810]: I0930 08:03:15.453054 4810 scope.go:117] "RemoveContainer" containerID="a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54" Sep 30 08:03:15 crc kubenswrapper[4810]: E0930 08:03:15.453405 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 08:03:15 crc kubenswrapper[4810]: I0930 08:03:15.750745 4810 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Sep 30 08:03:15 crc kubenswrapper[4810]: I0930 08:03:15.782498 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.452859 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.454209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.454310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.454343 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.455250 4810 scope.go:117] "RemoveContainer" containerID="a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54" Sep 30 08:03:16 crc kubenswrapper[4810]: E0930 08:03:16.455672 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.461198 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.798254 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:03:16 crc kubenswrapper[4810]: E0930 08:03:16.896259 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.900252 4810 trace.go:236] Trace[1604733413]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 08:03:06.373) (total time: 10526ms): Sep 30 08:03:16 crc kubenswrapper[4810]: Trace[1604733413]: ---"Objects listed" error: 10526ms (08:03:16.900) Sep 30 08:03:16 crc kubenswrapper[4810]: Trace[1604733413]: [10.526571199s] [10.526571199s] END Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.900328 4810 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.902003 4810 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.904524 4810 trace.go:236] Trace[2134379680]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 08:03:04.534) (total time: 12369ms): Sep 30 08:03:16 crc kubenswrapper[4810]: Trace[2134379680]: ---"Objects listed" error: 12369ms (08:03:16.904) Sep 30 08:03:16 crc kubenswrapper[4810]: Trace[2134379680]: [12.369690922s] [12.369690922s] END Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.904575 4810 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.908096 4810 trace.go:236] Trace[1540714134]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 08:03:06.197) (total time: 10710ms): Sep 30 08:03:16 crc kubenswrapper[4810]: Trace[1540714134]: ---"Objects listed" error: 10710ms (08:03:16.907) Sep 30 08:03:16 crc kubenswrapper[4810]: Trace[1540714134]: [10.710434772s] [10.710434772s] END Sep 30 08:03:16 crc kubenswrapper[4810]: I0930 08:03:16.908133 4810 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Sep 30 08:03:16 crc kubenswrapper[4810]: E0930 08:03:16.912173 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.224712 4810 apiserver.go:52] "Watching apiserver" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.228220 4810 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.228647 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.232365 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.232646 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.232721 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.232796 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.232916 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.234777 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.234835 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.234864 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.234793 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.238010 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.239434 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.251154 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.251653 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.252879 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.253021 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.253059 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.253125 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.253447 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.293706 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.304768 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.304819 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.304840 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.306153 4810 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.312519 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.320359 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.329083 4810 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.349794 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.382481 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405080 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405120 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405139 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405156 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405177 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405193 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405210 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405227 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405248 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405284 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405299 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405316 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405332 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405347 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405365 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405383 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405399 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405416 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405431 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405460 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405475 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405489 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405503 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405517 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405532 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405557 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405570 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405593 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405610 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405623 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405638 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405654 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405669 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405685 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405705 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405719 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405735 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405749 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405765 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405795 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405811 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405825 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405841 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405857 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405872 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405886 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405901 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405916 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405931 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405946 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405951 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405962 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405977 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.405992 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406007 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406022 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406035 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406050 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406064 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406078 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406093 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406112 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406127 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406143 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406158 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406173 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406188 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406202 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406217 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406233 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406250 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406278 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406293 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406308 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406322 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406339 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406353 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406370 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406411 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406428 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406442 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406457 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406473 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406487 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406504 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406519 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406533 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406547 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406561 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406577 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406594 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406609 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406625 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406642 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406657 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406672 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406689 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406705 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406720 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406748 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406765 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406783 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406799 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406816 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406831 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406847 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406863 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406879 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406895 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406910 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406925 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406941 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406956 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406974 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406991 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407007 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407022 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407038 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407054 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407069 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407084 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407100 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407117 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407132 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407147 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407163 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407179 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407193 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407208 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407224 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407239 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407255 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407284 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407299 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407314 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407331 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407347 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407362 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407381 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407396 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407411 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407427 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407441 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407457 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407473 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407487 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407502 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407518 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407534 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407549 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407565 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407605 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406092 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406210 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406688 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.406817 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407213 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.407531 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.412768 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.412831 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.412932 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.413015 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.413125 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.413253 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.413469 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.413600 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.413722 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.413846 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.413940 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.413985 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.414086 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.414150 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.414845 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.414879 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.415056 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.415544 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.415593 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.415822 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.416141 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.416407 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.416412 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.416569 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.416714 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.416741 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.417095 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.417131 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.417292 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.417395 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.417540 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.417692 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.417705 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.417743 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.417790 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.417826 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.417958 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.419179 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.418097 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.418139 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.418225 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.418501 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.418689 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.418685 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.418045 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.419681 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.419936 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.419985 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.420167 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.420426 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.420506 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.420732 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.428362 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.428407 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.428740 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.428790 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429001 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429207 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429420 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429442 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429547 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429752 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429828 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429833 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429847 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429971 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.430142 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.430311 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.430629 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.430712 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.430992 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431028 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431054 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431064 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431087 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431107 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431124 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431141 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431159 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431177 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431185 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431201 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431196 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431221 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431231 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431249 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431239 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431405 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431406 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431422 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431450 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431468 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431505 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431540 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431554 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431577 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431612 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431644 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431677 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431695 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431704 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431711 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431746 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431764 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431773 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431788 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431849 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431864 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431883 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431908 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431934 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431955 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431962 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.431975 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432050 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432070 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432086 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432103 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432122 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432139 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432157 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432172 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432188 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432203 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432220 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432238 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432254 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432562 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432562 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432582 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432599 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432607 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432642 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432674 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432693 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432714 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432918 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432937 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.432948 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433151 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433621 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433652 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433674 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433700 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433723 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433741 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433765 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433783 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433861 4810 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433874 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433888 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433900 4810 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433910 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433920 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433930 4810 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433939 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433949 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433959 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440369 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440460 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440522 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440584 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440643 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440700 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440754 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440810 4810 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440866 4810 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440933 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440986 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441041 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441092 4810 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441142 4810 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441216 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441283 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441344 4810 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441409 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441464 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441521 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441584 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441636 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441688 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441743 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441797 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441852 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441907 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.441961 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442011 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442064 4810 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442115 4810 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442171 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442229 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442413 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442473 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442532 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442584 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442642 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442703 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442761 4810 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442819 4810 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442872 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442930 4810 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.442981 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443039 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443103 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443168 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443235 4810 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443314 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443371 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443426 4810 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443477 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443535 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443587 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443714 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443771 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443822 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443875 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443929 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.443985 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.444036 4810 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.444094 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.444157 4810 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.444207 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.444260 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.445526 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.445596 4810 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.445658 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.445717 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.445768 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.445818 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.445872 4810 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.445928 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.445983 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446036 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446094 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446152 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446219 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446306 4810 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446378 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446437 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446492 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446547 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446600 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446654 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446707 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446757 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.446837 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.433875 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.434045 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.418325 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.434061 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.434377 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.434497 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.434734 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.434885 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.435088 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.435014 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.435321 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.435381 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.435399 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.435406 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.435487 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.435880 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.435928 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.436197 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.436197 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.436213 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.436222 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.436280 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.436545 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.436568 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429670 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.429984 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.436798 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.437117 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.437227 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.440328 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.447385 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.447737 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.448341 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.448500 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.448560 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.448642 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.448724 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.448423 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.449170 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.450038 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.450157 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.450388 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.450590 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.450851 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.451104 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.451179 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.451351 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.451448 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.451585 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.451596 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.451775 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.452010 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.452056 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.452329 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.452405 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.452461 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:03:17.95243081 +0000 UTC m=+21.404630077 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.452551 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.452592 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.452693 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.453048 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.455929 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.456021 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:17.956001003 +0000 UTC m=+21.408200280 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.456078 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.456814 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.456989 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.457255 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.458740 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.459779 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:17.959763522 +0000 UTC m=+21.411962779 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.459779 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.459813 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.458852 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.459958 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.460063 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.460374 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.460630 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.461909 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.462185 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.462387 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.462457 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.462919 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.463473 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.468407 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.469962 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.471462 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.472656 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.473797 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.479355 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.479543 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.479750 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.489567 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.490462 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.490793 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.490979 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.491005 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.491020 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.491081 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:17.991061205 +0000 UTC m=+21.443260492 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.491541 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.491567 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.491715 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.492455 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.493648 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.494177 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.494197 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.494194 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.494212 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.494350 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:17.99432635 +0000 UTC m=+21.446525717 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.494573 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.494622 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.494687 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.495173 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.499017 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.500493 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.500733 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.501094 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.503383 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.504976 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.509527 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.509925 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.516247 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.526956 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.534103 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.542389 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548087 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548162 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548174 4810 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548183 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548192 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548199 4810 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548207 4810 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548215 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548224 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548232 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548241 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548249 4810 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548258 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548282 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548290 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548298 4810 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548306 4810 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548314 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548323 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548340 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548348 4810 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548355 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548363 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548372 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548380 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548388 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548396 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548403 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548411 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548419 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548426 4810 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548436 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548444 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548452 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548459 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548469 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548477 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548485 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548493 4810 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548502 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548511 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548519 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548528 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548537 4810 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548545 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548553 4810 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548570 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548581 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548592 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548602 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548612 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548625 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548634 4810 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548644 4810 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548653 4810 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548661 4810 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548669 4810 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548679 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548687 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548695 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548705 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548714 4810 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548723 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548731 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548738 4810 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548747 4810 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548755 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548766 4810 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548774 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548783 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548792 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548802 4810 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548810 4810 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548818 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548826 4810 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548834 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548843 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548851 4810 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548859 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548867 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548876 4810 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548884 4810 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548894 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548901 4810 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548909 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548918 4810 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548926 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548927 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.548935 4810 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.549010 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.549025 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.549040 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.549053 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.549066 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.549078 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.549103 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.549115 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.549128 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.554088 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.556371 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.565536 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.567958 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.571808 4810 scope.go:117] "RemoveContainer" containerID="a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54" Sep 30 08:03:17 crc kubenswrapper[4810]: E0930 08:03:17.571967 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.572456 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.573212 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 08:03:17 crc kubenswrapper[4810]: I0930 08:03:17.578214 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:17 crc kubenswrapper[4810]: W0930 08:03:17.580036 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-f45f891e29e961e926e347ca53a7076c1f86c97898321222dfe1d8fda09eddc7 WatchSource:0}: Error finding container f45f891e29e961e926e347ca53a7076c1f86c97898321222dfe1d8fda09eddc7: Status 404 returned error can't find the container with id f45f891e29e961e926e347ca53a7076c1f86c97898321222dfe1d8fda09eddc7 Sep 30 08:03:17 crc kubenswrapper[4810]: W0930 08:03:17.581000 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-f65ceec8bc53f337e496518fa2cb02e985b83e0da0d855af806c9d927f0284c8 WatchSource:0}: Error finding container f65ceec8bc53f337e496518fa2cb02e985b83e0da0d855af806c9d927f0284c8: Status 404 returned error can't find the container with id f65ceec8bc53f337e496518fa2cb02e985b83e0da0d855af806c9d927f0284c8 Sep 30 08:03:17 crc kubenswrapper[4810]: W0930 08:03:17.592544 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-7717d9b650acf9ec378cd10fe644f1d7521012591c0c1514e3eec0d4a14be10c WatchSource:0}: Error finding container 7717d9b650acf9ec378cd10fe644f1d7521012591c0c1514e3eec0d4a14be10c: Status 404 returned error can't find the container with id 7717d9b650acf9ec378cd10fe644f1d7521012591c0c1514e3eec0d4a14be10c Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.052932 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.053052 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.053090 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.053123 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053154 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:03:19.053120714 +0000 UTC m=+22.505319981 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.053218 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053300 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053325 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053342 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053408 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:19.053386282 +0000 UTC m=+22.505585559 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053423 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053472 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:19.053464154 +0000 UTC m=+22.505663421 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053544 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053557 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053569 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053590 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:19.053584278 +0000 UTC m=+22.505783545 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053662 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.053687 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:19.05367903 +0000 UTC m=+22.505878297 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.275899 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-xrbr6"] Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.276235 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xrbr6" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.276385 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-n85f4"] Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.276790 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.277771 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.277789 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.278105 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.280865 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.280875 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.280875 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.281004 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.281189 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.281201 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-hk2k2"] Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.281712 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.283060 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-fnzgz"] Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.283646 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.285487 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.286189 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.286356 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.286550 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.287562 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.287702 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.287910 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.294970 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.305326 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.305446 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.313158 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.335855 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.344105 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356026 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-run-k8s-cni-cncf-io\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356068 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-var-lib-cni-bin\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356085 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-cnibin\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356099 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsl8p\" (UniqueName: \"kubernetes.io/projected/f5367714-1633-4695-af4e-4eb6419daa96-kube-api-access-qsl8p\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356119 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/128f6072-cce3-4aa1-beae-baa0a4f7d03b-cni-binary-copy\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356134 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-multus-cni-dir\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356148 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-multus-conf-dir\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356164 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-run-multus-certs\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356177 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-system-cni-dir\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356190 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-var-lib-kubelet\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356205 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-cnibin\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356219 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-os-release\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356232 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-hostroot\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356245 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wknsz\" (UniqueName: \"kubernetes.io/projected/128f6072-cce3-4aa1-beae-baa0a4f7d03b-kube-api-access-wknsz\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356282 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/86194959-30fe-4da8-a64b-f5a849c2f73a-rootfs\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356297 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffdsb\" (UniqueName: \"kubernetes.io/projected/86194959-30fe-4da8-a64b-f5a849c2f73a-kube-api-access-ffdsb\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356311 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f5367714-1633-4695-af4e-4eb6419daa96-cni-binary-copy\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356324 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-etc-kubernetes\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356350 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kvgm\" (UniqueName: \"kubernetes.io/projected/ec5a0024-ccdd-42f8-b7ed-1e395209d05b-kube-api-access-9kvgm\") pod \"node-resolver-xrbr6\" (UID: \"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\") " pod="openshift-dns/node-resolver-xrbr6" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356366 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/128f6072-cce3-4aa1-beae-baa0a4f7d03b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356382 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/86194959-30fe-4da8-a64b-f5a849c2f73a-mcd-auth-proxy-config\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356395 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-os-release\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356407 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-var-lib-cni-multus\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356421 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356444 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-run-netns\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356458 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/86194959-30fe-4da8-a64b-f5a849c2f73a-proxy-tls\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356472 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-system-cni-dir\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356489 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ec5a0024-ccdd-42f8-b7ed-1e395209d05b-hosts-file\") pod \"node-resolver-xrbr6\" (UID: \"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\") " pod="openshift-dns/node-resolver-xrbr6" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356518 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-multus-socket-dir-parent\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.356537 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f5367714-1633-4695-af4e-4eb6419daa96-multus-daemon-config\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.357848 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.373057 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.386399 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.399778 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.413772 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.427694 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.445314 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457255 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/86194959-30fe-4da8-a64b-f5a849c2f73a-rootfs\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457334 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffdsb\" (UniqueName: \"kubernetes.io/projected/86194959-30fe-4da8-a64b-f5a849c2f73a-kube-api-access-ffdsb\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457340 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/86194959-30fe-4da8-a64b-f5a849c2f73a-rootfs\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457377 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f5367714-1633-4695-af4e-4eb6419daa96-cni-binary-copy\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457400 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-etc-kubernetes\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457421 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kvgm\" (UniqueName: \"kubernetes.io/projected/ec5a0024-ccdd-42f8-b7ed-1e395209d05b-kube-api-access-9kvgm\") pod \"node-resolver-xrbr6\" (UID: \"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\") " pod="openshift-dns/node-resolver-xrbr6" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/128f6072-cce3-4aa1-beae-baa0a4f7d03b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457529 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-etc-kubernetes\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457549 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/86194959-30fe-4da8-a64b-f5a849c2f73a-mcd-auth-proxy-config\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457626 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-os-release\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457648 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-var-lib-cni-multus\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457671 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457697 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-run-netns\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457714 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/86194959-30fe-4da8-a64b-f5a849c2f73a-proxy-tls\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457736 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-system-cni-dir\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457755 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ec5a0024-ccdd-42f8-b7ed-1e395209d05b-hosts-file\") pod \"node-resolver-xrbr6\" (UID: \"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\") " pod="openshift-dns/node-resolver-xrbr6" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457775 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-multus-socket-dir-parent\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457789 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f5367714-1633-4695-af4e-4eb6419daa96-multus-daemon-config\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457815 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-run-k8s-cni-cncf-io\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457816 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-os-release\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457851 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-var-lib-cni-bin\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457829 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-var-lib-cni-bin\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457876 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-system-cni-dir\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457897 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-var-lib-cni-multus\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457911 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-cnibin\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457933 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-run-netns\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457942 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsl8p\" (UniqueName: \"kubernetes.io/projected/f5367714-1633-4695-af4e-4eb6419daa96-kube-api-access-qsl8p\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457944 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-cnibin\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457899 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ec5a0024-ccdd-42f8-b7ed-1e395209d05b-hosts-file\") pod \"node-resolver-xrbr6\" (UID: \"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\") " pod="openshift-dns/node-resolver-xrbr6" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.457999 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/128f6072-cce3-4aa1-beae-baa0a4f7d03b-cni-binary-copy\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458018 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-multus-cni-dir\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458034 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-multus-socket-dir-parent\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458055 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-multus-conf-dir\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458071 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-run-multus-certs\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458085 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-system-cni-dir\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458107 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-run-k8s-cni-cncf-io\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458222 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-multus-cni-dir\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458258 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-run-multus-certs\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458308 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-multus-conf-dir\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458330 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-system-cni-dir\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458322 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458367 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-var-lib-kubelet\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458402 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-cnibin\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458420 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-os-release\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458433 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-host-var-lib-kubelet\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458453 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-cnibin\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458460 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-hostroot\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458437 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f5367714-1633-4695-af4e-4eb6419daa96-hostroot\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458494 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wknsz\" (UniqueName: \"kubernetes.io/projected/128f6072-cce3-4aa1-beae-baa0a4f7d03b-kube-api-access-wknsz\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458505 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/128f6072-cce3-4aa1-beae-baa0a4f7d03b-os-release\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458684 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/128f6072-cce3-4aa1-beae-baa0a4f7d03b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458763 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f5367714-1633-4695-af4e-4eb6419daa96-cni-binary-copy\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458791 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f5367714-1633-4695-af4e-4eb6419daa96-multus-daemon-config\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.458845 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/128f6072-cce3-4aa1-beae-baa0a4f7d03b-cni-binary-copy\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.459042 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/86194959-30fe-4da8-a64b-f5a849c2f73a-mcd-auth-proxy-config\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.462604 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/86194959-30fe-4da8-a64b-f5a849c2f73a-proxy-tls\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.467821 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.468913 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0"} Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.468952 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8"} Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.468962 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7717d9b650acf9ec378cd10fe644f1d7521012591c0c1514e3eec0d4a14be10c"} Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.469896 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f65ceec8bc53f337e496518fa2cb02e985b83e0da0d855af806c9d927f0284c8"} Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.471073 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563"} Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.471144 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f45f891e29e961e926e347ca53a7076c1f86c97898321222dfe1d8fda09eddc7"} Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.475865 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsl8p\" (UniqueName: \"kubernetes.io/projected/f5367714-1633-4695-af4e-4eb6419daa96-kube-api-access-qsl8p\") pod \"multus-n85f4\" (UID: \"f5367714-1633-4695-af4e-4eb6419daa96\") " pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.479950 4810 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.481251 4810 scope.go:117] "RemoveContainer" containerID="a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54" Sep 30 08:03:18 crc kubenswrapper[4810]: E0930 08:03:18.481517 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.482744 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffdsb\" (UniqueName: \"kubernetes.io/projected/86194959-30fe-4da8-a64b-f5a849c2f73a-kube-api-access-ffdsb\") pod \"machine-config-daemon-hk2k2\" (UID: \"86194959-30fe-4da8-a64b-f5a849c2f73a\") " pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.483926 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wknsz\" (UniqueName: \"kubernetes.io/projected/128f6072-cce3-4aa1-beae-baa0a4f7d03b-kube-api-access-wknsz\") pod \"multus-additional-cni-plugins-fnzgz\" (UID: \"128f6072-cce3-4aa1-beae-baa0a4f7d03b\") " pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.484040 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kvgm\" (UniqueName: \"kubernetes.io/projected/ec5a0024-ccdd-42f8-b7ed-1e395209d05b-kube-api-access-9kvgm\") pod \"node-resolver-xrbr6\" (UID: \"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\") " pod="openshift-dns/node-resolver-xrbr6" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.504760 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.521279 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.536254 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.552198 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.566160 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.579774 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.591017 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xrbr6" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.595816 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.599350 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-n85f4" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.615650 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.622210 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.636603 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.646768 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.660254 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.660687 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9bkcj"] Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.661501 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.664321 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.664591 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.664743 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.665130 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.665173 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.665277 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.665683 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.680015 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.698812 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.710746 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.725110 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.739053 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.751139 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761424 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-ovn\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761460 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovn-node-metrics-cert\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761476 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-script-lib\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761492 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-node-log\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761515 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-slash\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761597 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-ovn-kubernetes\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761651 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-netd\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761710 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761764 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-etc-openvswitch\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761798 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-netns\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761827 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-env-overrides\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761848 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-systemd\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761876 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-systemd-units\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761891 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-bin\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761928 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-kubelet\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761955 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-config\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.761987 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-openvswitch\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.762011 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-log-socket\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.762054 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-var-lib-openvswitch\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.762087 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx6mf\" (UniqueName: \"kubernetes.io/projected/aacbcdc8-7ef5-473b-8055-145f40040d5d-kube-api-access-kx6mf\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.766189 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.783982 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.801568 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.825476 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.853461 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862606 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-kubelet\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862650 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-config\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862686 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-openvswitch\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862708 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-log-socket\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862730 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-var-lib-openvswitch\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862758 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx6mf\" (UniqueName: \"kubernetes.io/projected/aacbcdc8-7ef5-473b-8055-145f40040d5d-kube-api-access-kx6mf\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862777 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovn-node-metrics-cert\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862798 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-script-lib\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862822 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-ovn\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862834 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-log-socket\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862850 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-node-log\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862871 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-slash\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862891 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-ovn-kubernetes\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862912 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862938 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-netd\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862967 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-etc-openvswitch\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862996 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-netns\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.863018 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-env-overrides\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.863040 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-systemd\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.863060 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-bin\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.863080 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-systemd-units\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.863155 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-systemd-units\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.863205 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-var-lib-openvswitch\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.863756 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-config\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.863828 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.864144 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-netns\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.864204 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-netd\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.864230 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-etc-openvswitch\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.864256 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-openvswitch\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.864294 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-node-log\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.862787 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-kubelet\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.864327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-ovn\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.864362 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-systemd\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.864398 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-script-lib\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.864451 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-slash\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.864488 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-ovn-kubernetes\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.864527 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-bin\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.865880 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-env-overrides\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.870451 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.870766 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovn-node-metrics-cert\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.884564 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx6mf\" (UniqueName: \"kubernetes.io/projected/aacbcdc8-7ef5-473b-8055-145f40040d5d-kube-api-access-kx6mf\") pod \"ovnkube-node-9bkcj\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.886641 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.903931 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.920857 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.934993 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.954286 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.971314 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.971471 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.986408 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:18 crc kubenswrapper[4810]: W0930 08:03:18.997088 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaacbcdc8_7ef5_473b_8055_145f40040d5d.slice/crio-45d54bd719d7b34f0b30fe78e45e73f656954ceb4b9e69c6dba27b382d9ee125 WatchSource:0}: Error finding container 45d54bd719d7b34f0b30fe78e45e73f656954ceb4b9e69c6dba27b382d9ee125: Status 404 returned error can't find the container with id 45d54bd719d7b34f0b30fe78e45e73f656954ceb4b9e69c6dba27b382d9ee125 Sep 30 08:03:18 crc kubenswrapper[4810]: I0930 08:03:18.997211 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:18Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.065072 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.065250 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.065354 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.065427 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.065504 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.065661 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.065726 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.065778 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.065861 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:21.065847496 +0000 UTC m=+24.518046763 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.066178 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:03:21.066168325 +0000 UTC m=+24.518367592 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.066297 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.066371 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:21.066363671 +0000 UTC m=+24.518562938 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.066462 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.066533 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.066587 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.066672 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:21.06666507 +0000 UTC m=+24.518864337 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.066791 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.066872 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:21.066864906 +0000 UTC m=+24.519064173 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.306025 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.306347 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.306093 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.306826 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.309802 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.310331 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.311107 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.311759 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.312403 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.312913 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.313557 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.314078 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.316096 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.316940 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.318248 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.318915 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.319803 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.320434 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.321291 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.321778 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.322340 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.325639 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.326255 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.326832 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.327716 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.328351 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.329148 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.329794 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.330226 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.331217 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.332319 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.332773 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.333359 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.334188 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.334655 4810 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.334757 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.337703 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.338238 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.338708 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.340687 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.341613 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.342135 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.343104 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.343738 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.344563 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.345120 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.346095 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.347031 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.347722 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.351465 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.352076 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.353171 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.353679 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.354139 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.354978 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.355506 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.356445 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.356887 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.475048 4810 generic.go:334] "Generic (PLEG): container finished" podID="128f6072-cce3-4aa1-beae-baa0a4f7d03b" containerID="21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f" exitCode=0 Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.475140 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" event={"ID":"128f6072-cce3-4aa1-beae-baa0a4f7d03b","Type":"ContainerDied","Data":"21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f"} Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.476242 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" event={"ID":"128f6072-cce3-4aa1-beae-baa0a4f7d03b","Type":"ContainerStarted","Data":"0f5fe7a3d3b2e812377bbc4bb8f338c76ed3d2ad6f06f7f83a9b3ef70e434b07"} Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.477649 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479" exitCode=0 Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.477734 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479"} Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.477765 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"45d54bd719d7b34f0b30fe78e45e73f656954ceb4b9e69c6dba27b382d9ee125"} Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.479710 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd"} Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.479745 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4"} Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.479759 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"97386a9961f8c183f7d2e5d7ec9782b705fe946e8b8b60bfeb222cc9edaef5e9"} Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.481200 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n85f4" event={"ID":"f5367714-1633-4695-af4e-4eb6419daa96","Type":"ContainerStarted","Data":"de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab"} Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.481240 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n85f4" event={"ID":"f5367714-1633-4695-af4e-4eb6419daa96","Type":"ContainerStarted","Data":"a6117655f5135357764d36690bbfc427b1f30692b9be30335a158f4a58c63dd0"} Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.483082 4810 scope.go:117] "RemoveContainer" containerID="a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54" Sep 30 08:03:19 crc kubenswrapper[4810]: E0930 08:03:19.483201 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.483234 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xrbr6" event={"ID":"ec5a0024-ccdd-42f8-b7ed-1e395209d05b","Type":"ContainerStarted","Data":"254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30"} Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.483250 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xrbr6" event={"ID":"ec5a0024-ccdd-42f8-b7ed-1e395209d05b","Type":"ContainerStarted","Data":"75a1ad34286a3306187ab5d44e1e167bc5686f75b727abf06f1fb73daddce43d"} Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.502675 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.516549 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.548985 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.568846 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.582812 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.596178 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.610477 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.633279 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.649889 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.663866 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.679957 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.689395 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.704378 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.722440 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.741201 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.755989 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.772009 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.789093 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.805348 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.817150 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.829748 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.849100 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.868663 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:19 crc kubenswrapper[4810]: I0930 08:03:19.881923 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:19Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.154458 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.176823 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.191389 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.191860 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.215600 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.244084 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.266895 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.281772 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.293545 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.306303 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:20 crc kubenswrapper[4810]: E0930 08:03:20.306416 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.306760 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.322754 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.336879 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.347741 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.359852 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.372298 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.396285 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.419992 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.439222 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.454059 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.467585 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.487430 4810 generic.go:334] "Generic (PLEG): container finished" podID="128f6072-cce3-4aa1-beae-baa0a4f7d03b" containerID="acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b" exitCode=0 Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.487472 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.487539 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" event={"ID":"128f6072-cce3-4aa1-beae-baa0a4f7d03b","Type":"ContainerDied","Data":"acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b"} Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.488627 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c"} Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.491842 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38"} Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.491898 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5"} Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.491915 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5"} Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.491929 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644"} Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.504016 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.519367 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.533925 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.551640 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.573380 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.584124 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.606699 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.623353 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.637492 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.650551 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.668864 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.711407 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.748533 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.784977 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.827452 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.871170 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.912791 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.947538 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:20 crc kubenswrapper[4810]: I0930 08:03:20.986605 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:20Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.034243 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.085913 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086144 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:03:25.086103289 +0000 UTC m=+28.538302586 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.086225 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.086363 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.086424 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086437 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.086547 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086573 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:25.086540941 +0000 UTC m=+28.538740238 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086720 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086769 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086777 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086796 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086815 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086840 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086864 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:25.08684517 +0000 UTC m=+28.539044467 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086775 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086917 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:25.086894801 +0000 UTC m=+28.539094168 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.086952 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:25.086935323 +0000 UTC m=+28.539134770 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.305845 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.305936 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.306126 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:21 crc kubenswrapper[4810]: E0930 08:03:21.306300 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.333717 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-z5md6"] Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.334239 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-z5md6" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.336882 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.337084 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.337556 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.337667 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.358038 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.375570 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.389714 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.389843 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/154f9139-cbed-4ce1-a15b-b28a19016fab-host\") pod \"node-ca-z5md6\" (UID: \"154f9139-cbed-4ce1-a15b-b28a19016fab\") " pod="openshift-image-registry/node-ca-z5md6" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.389870 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtts8\" (UniqueName: \"kubernetes.io/projected/154f9139-cbed-4ce1-a15b-b28a19016fab-kube-api-access-wtts8\") pod \"node-ca-z5md6\" (UID: \"154f9139-cbed-4ce1-a15b-b28a19016fab\") " pod="openshift-image-registry/node-ca-z5md6" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.389889 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/154f9139-cbed-4ce1-a15b-b28a19016fab-serviceca\") pod \"node-ca-z5md6\" (UID: \"154f9139-cbed-4ce1-a15b-b28a19016fab\") " pod="openshift-image-registry/node-ca-z5md6" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.399721 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.410426 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.425444 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.447917 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.464126 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.475224 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.491217 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/154f9139-cbed-4ce1-a15b-b28a19016fab-host\") pod \"node-ca-z5md6\" (UID: \"154f9139-cbed-4ce1-a15b-b28a19016fab\") " pod="openshift-image-registry/node-ca-z5md6" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.491280 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtts8\" (UniqueName: \"kubernetes.io/projected/154f9139-cbed-4ce1-a15b-b28a19016fab-kube-api-access-wtts8\") pod \"node-ca-z5md6\" (UID: \"154f9139-cbed-4ce1-a15b-b28a19016fab\") " pod="openshift-image-registry/node-ca-z5md6" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.491302 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/154f9139-cbed-4ce1-a15b-b28a19016fab-serviceca\") pod \"node-ca-z5md6\" (UID: \"154f9139-cbed-4ce1-a15b-b28a19016fab\") " pod="openshift-image-registry/node-ca-z5md6" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.491334 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/154f9139-cbed-4ce1-a15b-b28a19016fab-host\") pod \"node-ca-z5md6\" (UID: \"154f9139-cbed-4ce1-a15b-b28a19016fab\") " pod="openshift-image-registry/node-ca-z5md6" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.494091 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/154f9139-cbed-4ce1-a15b-b28a19016fab-serviceca\") pod \"node-ca-z5md6\" (UID: \"154f9139-cbed-4ce1-a15b-b28a19016fab\") " pod="openshift-image-registry/node-ca-z5md6" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.498991 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89"} Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.499031 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f"} Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.501186 4810 generic.go:334] "Generic (PLEG): container finished" podID="128f6072-cce3-4aa1-beae-baa0a4f7d03b" containerID="45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529" exitCode=0 Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.501313 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" event={"ID":"128f6072-cce3-4aa1-beae-baa0a4f7d03b","Type":"ContainerDied","Data":"45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529"} Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.512846 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.538605 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtts8\" (UniqueName: \"kubernetes.io/projected/154f9139-cbed-4ce1-a15b-b28a19016fab-kube-api-access-wtts8\") pod \"node-ca-z5md6\" (UID: \"154f9139-cbed-4ce1-a15b-b28a19016fab\") " pod="openshift-image-registry/node-ca-z5md6" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.567874 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.609604 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.648328 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.659544 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-z5md6" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.685538 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.729336 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.768533 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.807339 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.845367 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.888088 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.927514 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:21 crc kubenswrapper[4810]: I0930 08:03:21.970487 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:21Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.006487 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.048960 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.099582 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.127392 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.131462 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.132990 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.149431 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.216749 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.231225 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.266419 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.306611 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:22 crc kubenswrapper[4810]: E0930 08:03:22.306802 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.315548 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.346376 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.390274 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.428327 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.470486 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.507861 4810 generic.go:334] "Generic (PLEG): container finished" podID="128f6072-cce3-4aa1-beae-baa0a4f7d03b" containerID="a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d" exitCode=0 Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.507908 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" event={"ID":"128f6072-cce3-4aa1-beae-baa0a4f7d03b","Type":"ContainerDied","Data":"a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d"} Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.510120 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-z5md6" event={"ID":"154f9139-cbed-4ce1-a15b-b28a19016fab","Type":"ContainerStarted","Data":"5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a"} Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.510386 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-z5md6" event={"ID":"154f9139-cbed-4ce1-a15b-b28a19016fab","Type":"ContainerStarted","Data":"a23ca5f4d1e8525278f4b1dca1e95cd0798fb4c20c9759529edf2c15da4c2e97"} Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.514306 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.560094 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.592485 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.629381 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.666557 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.714044 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.748176 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.786799 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.832829 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.868019 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.916064 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.948441 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:22 crc kubenswrapper[4810]: I0930 08:03:22.991985 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:22Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.028481 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.070366 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.106344 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.155194 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.188801 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.229210 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.269810 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.305613 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.305741 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:23 crc kubenswrapper[4810]: E0930 08:03:23.305809 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:23 crc kubenswrapper[4810]: E0930 08:03:23.305925 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.311118 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.312255 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.314148 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.314196 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.314214 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.314330 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.369735 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.379348 4810 kubelet_node_status.go:115] "Node was previously registered" node="crc" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.379717 4810 kubelet_node_status.go:79] "Successfully registered node" node="crc" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.381292 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.381331 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.381349 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.381374 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.381392 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:23Z","lastTransitionTime":"2025-09-30T08:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:23 crc kubenswrapper[4810]: E0930 08:03:23.396426 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.401093 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.401139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.401148 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.401166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.401178 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:23Z","lastTransitionTime":"2025-09-30T08:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:23 crc kubenswrapper[4810]: E0930 08:03:23.421223 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.425347 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.425374 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.425383 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.425406 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.425416 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:23Z","lastTransitionTime":"2025-09-30T08:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.433521 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: E0930 08:03:23.443139 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.447115 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.447154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.447167 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.447185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.447201 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:23Z","lastTransitionTime":"2025-09-30T08:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:23 crc kubenswrapper[4810]: E0930 08:03:23.460165 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.463667 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.463852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.463981 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.464133 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.464250 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:23Z","lastTransitionTime":"2025-09-30T08:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.474792 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: E0930 08:03:23.488853 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: E0930 08:03:23.489076 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.491031 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.491065 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.491078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.491097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.491110 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:23Z","lastTransitionTime":"2025-09-30T08:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.507028 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.516816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6"} Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.519925 4810 generic.go:334] "Generic (PLEG): container finished" podID="128f6072-cce3-4aa1-beae-baa0a4f7d03b" containerID="12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747" exitCode=0 Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.519955 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" event={"ID":"128f6072-cce3-4aa1-beae-baa0a4f7d03b","Type":"ContainerDied","Data":"12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747"} Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.549842 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.593685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.593836 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.593897 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.593960 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.594027 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:23Z","lastTransitionTime":"2025-09-30T08:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.607716 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.631147 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.671101 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.696886 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.696932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.696943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.696961 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.696976 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:23Z","lastTransitionTime":"2025-09-30T08:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.711488 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.751261 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.792562 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.800860 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.800905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.800918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.800937 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.800949 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:23Z","lastTransitionTime":"2025-09-30T08:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.840608 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.867688 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.903208 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.903319 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.903380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.903453 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.903508 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:23Z","lastTransitionTime":"2025-09-30T08:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.909737 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.950775 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:23 crc kubenswrapper[4810]: I0930 08:03:23.991978 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:23Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.005905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.005963 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.005982 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.006007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.006025 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:24Z","lastTransitionTime":"2025-09-30T08:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.032084 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.071752 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.107781 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.108145 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.108239 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.108313 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.108381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.108445 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:24Z","lastTransitionTime":"2025-09-30T08:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.210862 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.211367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.211523 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.211648 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.211816 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:24Z","lastTransitionTime":"2025-09-30T08:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.305499 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:24 crc kubenswrapper[4810]: E0930 08:03:24.305696 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.314068 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.314114 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.314132 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.314156 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.314175 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:24Z","lastTransitionTime":"2025-09-30T08:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.417322 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.417412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.417436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.417462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.417484 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:24Z","lastTransitionTime":"2025-09-30T08:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.519947 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.519998 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.520016 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.520043 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.520067 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:24Z","lastTransitionTime":"2025-09-30T08:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.530497 4810 generic.go:334] "Generic (PLEG): container finished" podID="128f6072-cce3-4aa1-beae-baa0a4f7d03b" containerID="87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891" exitCode=0 Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.530564 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" event={"ID":"128f6072-cce3-4aa1-beae-baa0a4f7d03b","Type":"ContainerDied","Data":"87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891"} Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.569341 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.599210 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.619401 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.623415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.623465 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.623483 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.623507 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.623525 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:24Z","lastTransitionTime":"2025-09-30T08:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.636204 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.655936 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.678563 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.701668 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.727159 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.727224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.727242 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.727291 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.727310 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:24Z","lastTransitionTime":"2025-09-30T08:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.727669 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.742758 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.765866 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.781073 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.796947 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.813374 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.827358 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.829493 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.829562 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.829587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.829618 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.829641 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:24Z","lastTransitionTime":"2025-09-30T08:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.843011 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:24Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.932486 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.932554 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.932571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.932599 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:24 crc kubenswrapper[4810]: I0930 08:03:24.932617 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:24Z","lastTransitionTime":"2025-09-30T08:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.035930 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.035994 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.036014 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.036040 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.036061 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:25Z","lastTransitionTime":"2025-09-30T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.131186 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.131314 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.131355 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.131382 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.131406 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131445 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:03:33.131407303 +0000 UTC m=+36.583606610 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131524 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131565 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131578 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:33.131564277 +0000 UTC m=+36.583763544 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131625 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131649 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131656 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:33.131639279 +0000 UTC m=+36.583838666 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131664 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131696 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:33.131688221 +0000 UTC m=+36.583887488 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131739 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131790 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131811 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.131888 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:33.131863056 +0000 UTC m=+36.584062363 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.138097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.138123 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.138135 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.138153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.138165 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:25Z","lastTransitionTime":"2025-09-30T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.241617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.241654 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.241664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.241678 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.241686 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:25Z","lastTransitionTime":"2025-09-30T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.306019 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.306184 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.306299 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:25 crc kubenswrapper[4810]: E0930 08:03:25.306405 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.347146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.347203 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.347221 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.347244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.347263 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:25Z","lastTransitionTime":"2025-09-30T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.450952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.451008 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.451027 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.451049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.451071 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:25Z","lastTransitionTime":"2025-09-30T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.538534 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" event={"ID":"128f6072-cce3-4aa1-beae-baa0a4f7d03b","Type":"ContainerStarted","Data":"bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.547695 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.548184 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.548210 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.553004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.553045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.553058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.553073 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.553086 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:25Z","lastTransitionTime":"2025-09-30T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.561667 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.582886 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.591427 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.594992 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.621987 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.635986 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.652799 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.655575 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.655619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.655636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.655660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.655677 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:25Z","lastTransitionTime":"2025-09-30T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.671482 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.684419 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.698046 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.712943 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.731272 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.758773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.758846 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.758863 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.758888 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.758906 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:25Z","lastTransitionTime":"2025-09-30T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.763365 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.783473 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.805536 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.824991 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.843122 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.856559 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.861150 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.861196 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.861213 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.861238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.861250 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:25Z","lastTransitionTime":"2025-09-30T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.869430 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.897871 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.912529 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.932128 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.950501 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.963604 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.963653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.963669 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.963692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.963708 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:25Z","lastTransitionTime":"2025-09-30T08:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:25 crc kubenswrapper[4810]: I0930 08:03:25.978930 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.017020 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:26Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.038584 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:26Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.065407 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:26Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.066117 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.066154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.066163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.066176 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.066185 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:26Z","lastTransitionTime":"2025-09-30T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.078942 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:26Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.093172 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:26Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.104897 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:26Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.117101 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:26Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.127928 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:26Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.170110 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.170149 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.170159 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.170175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.170186 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:26Z","lastTransitionTime":"2025-09-30T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.273647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.273692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.273705 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.273723 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.273735 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:26Z","lastTransitionTime":"2025-09-30T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.306563 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:26 crc kubenswrapper[4810]: E0930 08:03:26.306842 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.377432 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.377786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.377809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.377841 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.377863 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:26Z","lastTransitionTime":"2025-09-30T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.481328 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.481405 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.481426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.481462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.481487 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:26Z","lastTransitionTime":"2025-09-30T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.560495 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.585189 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.585242 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.585259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.585313 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.585332 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:26Z","lastTransitionTime":"2025-09-30T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.689107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.689162 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.689179 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.689206 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.689223 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:26Z","lastTransitionTime":"2025-09-30T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.792149 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.792218 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.792237 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.792262 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.792316 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:26Z","lastTransitionTime":"2025-09-30T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.895896 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.895964 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.895981 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.896006 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:26 crc kubenswrapper[4810]: I0930 08:03:26.896025 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:26Z","lastTransitionTime":"2025-09-30T08:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.000178 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.000631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.000661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.000693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.000712 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:27Z","lastTransitionTime":"2025-09-30T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.103624 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.103692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.103711 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.103754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.103769 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:27Z","lastTransitionTime":"2025-09-30T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.206870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.206914 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.206926 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.206944 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.206955 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:27Z","lastTransitionTime":"2025-09-30T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.305676 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:27 crc kubenswrapper[4810]: E0930 08:03:27.305820 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.305871 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:27 crc kubenswrapper[4810]: E0930 08:03:27.305997 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.309668 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.309736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.309754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.309777 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.309795 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:27Z","lastTransitionTime":"2025-09-30T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.345576 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.362944 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.382171 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.399159 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.411845 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.411901 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.411920 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.411944 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.411965 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:27Z","lastTransitionTime":"2025-09-30T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.415627 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.436701 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.465775 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.488128 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.509911 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.514489 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.514534 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.514544 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.514560 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.514569 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:27Z","lastTransitionTime":"2025-09-30T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.525384 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.544604 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.564567 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.566138 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.579369 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.592159 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.604467 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.617304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.617361 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.617380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.617407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.617428 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:27Z","lastTransitionTime":"2025-09-30T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.720838 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.720902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.720921 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.720946 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.720965 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:27Z","lastTransitionTime":"2025-09-30T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.823799 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.823855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.823873 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.823897 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.823916 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:27Z","lastTransitionTime":"2025-09-30T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.926890 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.926972 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.926997 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.927031 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:27 crc kubenswrapper[4810]: I0930 08:03:27.927055 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:27Z","lastTransitionTime":"2025-09-30T08:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.030025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.030086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.030102 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.030127 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.030145 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:28Z","lastTransitionTime":"2025-09-30T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.132625 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.132685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.132703 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.132728 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.132743 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:28Z","lastTransitionTime":"2025-09-30T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.236132 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.236192 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.236210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.236238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.236257 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:28Z","lastTransitionTime":"2025-09-30T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.306229 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:28 crc kubenswrapper[4810]: E0930 08:03:28.306523 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.339168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.339229 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.339249 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.339309 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.339327 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:28Z","lastTransitionTime":"2025-09-30T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.442348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.442390 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.442402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.442416 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.442426 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:28Z","lastTransitionTime":"2025-09-30T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.545451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.545512 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.545525 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.545560 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.545572 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:28Z","lastTransitionTime":"2025-09-30T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.570758 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/0.log" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.574482 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8" exitCode=1 Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.574549 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8"} Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.576073 4810 scope.go:117] "RemoveContainer" containerID="b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.597708 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.616643 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.648972 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.649021 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.649038 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.649068 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.649089 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:28Z","lastTransitionTime":"2025-09-30T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.654633 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.673878 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.705661 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 08:03:28.127246 6117 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:28.127302 6117 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:28.127307 6117 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:28.127317 6117 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:28.127335 6117 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 08:03:28.127346 6117 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 08:03:28.127351 6117 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 08:03:28.127364 6117 factory.go:656] Stopping watch factory\\\\nI0930 08:03:28.127374 6117 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:28.127367 6117 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 08:03:28.127387 6117 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:28.127410 6117 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 08:03:28.127409 6117 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:28.127416 6117 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 08:03:28.127425 6117 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:2910\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.728599 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.752112 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.752455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.752503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.752524 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.752554 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.752576 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:28Z","lastTransitionTime":"2025-09-30T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.773865 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.795347 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.816759 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.838549 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.855400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.855451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.855463 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.855480 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.855493 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:28Z","lastTransitionTime":"2025-09-30T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.864018 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.884703 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.904593 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.917173 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:28Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.964674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.965033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.965255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.965432 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:28 crc kubenswrapper[4810]: I0930 08:03:28.965593 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:28Z","lastTransitionTime":"2025-09-30T08:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.068165 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.068197 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.068205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.068219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.068228 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:29Z","lastTransitionTime":"2025-09-30T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.171210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.171284 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.171297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.171316 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.171329 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:29Z","lastTransitionTime":"2025-09-30T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.274368 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.275005 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.275142 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.275249 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.275347 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:29Z","lastTransitionTime":"2025-09-30T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.306263 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.306338 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:29 crc kubenswrapper[4810]: E0930 08:03:29.306420 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:29 crc kubenswrapper[4810]: E0930 08:03:29.306565 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.378378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.378430 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.378442 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.378458 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.378469 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:29Z","lastTransitionTime":"2025-09-30T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.481632 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.481996 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.482166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.482371 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.482528 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:29Z","lastTransitionTime":"2025-09-30T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.582148 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/0.log" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.585010 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.585060 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.585072 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.585092 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.585104 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:29Z","lastTransitionTime":"2025-09-30T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.586636 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8"} Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.586844 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.602504 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.615552 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.641762 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.654723 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.666680 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.686431 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.686903 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.686957 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.686976 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.687001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.687020 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:29Z","lastTransitionTime":"2025-09-30T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.706798 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.730340 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.760802 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 08:03:28.127246 6117 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:28.127302 6117 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:28.127307 6117 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:28.127317 6117 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:28.127335 6117 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 08:03:28.127346 6117 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 08:03:28.127351 6117 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 08:03:28.127364 6117 factory.go:656] Stopping watch factory\\\\nI0930 08:03:28.127374 6117 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:28.127367 6117 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 08:03:28.127387 6117 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:28.127410 6117 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 08:03:28.127409 6117 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:28.127416 6117 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 08:03:28.127425 6117 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:2910\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.784472 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.789468 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.789519 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.789538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.789564 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.789582 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:29Z","lastTransitionTime":"2025-09-30T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.806838 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.826350 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.848989 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.868419 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.884900 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:29Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.892489 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.892544 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.892566 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.892594 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.892613 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:29Z","lastTransitionTime":"2025-09-30T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.995997 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.996060 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.996078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.996104 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:29 crc kubenswrapper[4810]: I0930 08:03:29.996127 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:29Z","lastTransitionTime":"2025-09-30T08:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.100155 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.100214 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.100233 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.100263 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.100320 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:30Z","lastTransitionTime":"2025-09-30T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.204258 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.204362 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.204376 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.204399 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.204412 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:30Z","lastTransitionTime":"2025-09-30T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.305385 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:30 crc kubenswrapper[4810]: E0930 08:03:30.305515 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.307519 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.307581 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.307602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.307633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.307655 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:30Z","lastTransitionTime":"2025-09-30T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.410647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.410709 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.410727 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.410752 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.410769 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:30Z","lastTransitionTime":"2025-09-30T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.514182 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.514251 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.514296 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.514324 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.514343 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:30Z","lastTransitionTime":"2025-09-30T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.593159 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/1.log" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.594151 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/0.log" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.597871 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8" exitCode=1 Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.597927 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8"} Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.597980 4810 scope.go:117] "RemoveContainer" containerID="b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.599389 4810 scope.go:117] "RemoveContainer" containerID="ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8" Sep 30 08:03:30 crc kubenswrapper[4810]: E0930 08:03:30.599746 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.618827 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.619073 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.619192 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.619385 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.619513 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:30Z","lastTransitionTime":"2025-09-30T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.624760 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.644222 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.676038 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.695829 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.716673 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.721942 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.722141 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.722334 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.722556 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.722695 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:30Z","lastTransitionTime":"2025-09-30T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.737648 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.757870 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.768043 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85"] Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.768766 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:30 crc kubenswrapper[4810]: W0930 08:03:30.770746 4810 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert": failed to list *v1.Secret: secrets "ovn-control-plane-metrics-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Sep 30 08:03:30 crc kubenswrapper[4810]: E0930 08:03:30.770845 4810 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-control-plane-metrics-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.771133 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.781937 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.804755 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.826697 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.826760 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.826778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.826809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.826828 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:30Z","lastTransitionTime":"2025-09-30T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.831092 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.866024 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 08:03:28.127246 6117 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:28.127302 6117 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:28.127307 6117 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:28.127317 6117 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:28.127335 6117 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 08:03:28.127346 6117 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 08:03:28.127351 6117 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 08:03:28.127364 6117 factory.go:656] Stopping watch factory\\\\nI0930 08:03:28.127374 6117 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:28.127367 6117 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 08:03:28.127387 6117 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:28.127410 6117 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 08:03:28.127409 6117 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:28.127416 6117 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 08:03:28.127425 6117 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:2910\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:29Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.590716 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.591154 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:29.591222 6235 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 08:03:29.591236 6235 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 08:03:29.591256 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:29.591310 6235 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 08:03:29.591335 6235 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:29.591348 6235 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 08:03:29.591366 6235 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:29.591379 6235 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:29.591427 6235 factory.go:656] Stopping watch factory\\\\nI0930 08:03:29.591442 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:29.591451 6235 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:29.591466 6235 handler.go:208] Removed *v1.Node event handler 2\\\\nI09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.889091 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.893542 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.893757 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.893835 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkb6h\" (UniqueName: \"kubernetes.io/projected/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-kube-api-access-bkb6h\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.894076 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.910943 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.927401 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.929662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.929753 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.929773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.929801 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.929829 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:30Z","lastTransitionTime":"2025-09-30T08:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.946020 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.969798 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.991093 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:30Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.994842 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkb6h\" (UniqueName: \"kubernetes.io/projected/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-kube-api-access-bkb6h\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.994942 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.995026 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.995059 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.996339 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:30 crc kubenswrapper[4810]: I0930 08:03:30.996585 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.016369 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.028853 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkb6h\" (UniqueName: \"kubernetes.io/projected/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-kube-api-access-bkb6h\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.034765 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.034819 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.034837 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.034889 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.034915 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:31Z","lastTransitionTime":"2025-09-30T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.037663 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.058810 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.075563 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.111974 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.129398 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.139020 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.139081 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.139100 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.139125 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.139145 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:31Z","lastTransitionTime":"2025-09-30T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.149605 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.175744 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.197523 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.216889 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.242391 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.242443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.242458 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.242479 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.242493 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:31Z","lastTransitionTime":"2025-09-30T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.245525 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.259624 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.282210 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.305404 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:31 crc kubenswrapper[4810]: E0930 08:03:31.305568 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.305705 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.305574 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 08:03:28.127246 6117 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:28.127302 6117 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:28.127307 6117 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:28.127317 6117 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:28.127335 6117 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 08:03:28.127346 6117 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 08:03:28.127351 6117 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 08:03:28.127364 6117 factory.go:656] Stopping watch factory\\\\nI0930 08:03:28.127374 6117 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:28.127367 6117 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 08:03:28.127387 6117 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:28.127410 6117 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 08:03:28.127409 6117 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:28.127416 6117 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 08:03:28.127425 6117 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:2910\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:29Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.590716 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.591154 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:29.591222 6235 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 08:03:29.591236 6235 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 08:03:29.591256 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:29.591310 6235 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 08:03:29.591335 6235 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:29.591348 6235 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 08:03:29.591366 6235 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:29.591379 6235 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:29.591427 6235 factory.go:656] Stopping watch factory\\\\nI0930 08:03:29.591442 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:29.591451 6235 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:29.591466 6235 handler.go:208] Removed *v1.Node event handler 2\\\\nI09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:31Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:31 crc kubenswrapper[4810]: E0930 08:03:31.305872 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.345575 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.345736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.345836 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.345942 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.346019 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:31Z","lastTransitionTime":"2025-09-30T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.449761 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.449832 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.449853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.449878 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.449895 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:31Z","lastTransitionTime":"2025-09-30T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.553533 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.553588 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.553607 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.553630 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.553644 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:31Z","lastTransitionTime":"2025-09-30T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.604704 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/1.log" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.656240 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.656343 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.656362 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.656454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.656477 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:31Z","lastTransitionTime":"2025-09-30T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.759108 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.759197 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.759216 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.759237 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.759253 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:31Z","lastTransitionTime":"2025-09-30T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.862695 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.862770 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.862787 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.862813 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.862832 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:31Z","lastTransitionTime":"2025-09-30T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.966650 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.966737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.966761 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.966792 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:31 crc kubenswrapper[4810]: I0930 08:03:31.966814 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:31Z","lastTransitionTime":"2025-09-30T08:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:31 crc kubenswrapper[4810]: E0930 08:03:31.996056 4810 secret.go:188] Couldn't get secret openshift-ovn-kubernetes/ovn-control-plane-metrics-cert: failed to sync secret cache: timed out waiting for the condition Sep 30 08:03:31 crc kubenswrapper[4810]: E0930 08:03:31.996183 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-ovn-control-plane-metrics-cert podName:022a6cd0-9f4e-4ee1-a6c5-68568c7e4029 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:32.496154615 +0000 UTC m=+35.948353912 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovn-control-plane-metrics-cert" (UniqueName: "kubernetes.io/secret/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-ovn-control-plane-metrics-cert") pod "ovnkube-control-plane-749d76644c-jlk85" (UID: "022a6cd0-9f4e-4ee1-a6c5-68568c7e4029") : failed to sync secret cache: timed out waiting for the condition Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.069998 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.070049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.070061 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.070082 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.070096 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:32Z","lastTransitionTime":"2025-09-30T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.072230 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.173111 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.173169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.173186 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.173209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.173226 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:32Z","lastTransitionTime":"2025-09-30T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.276366 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.276424 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.276441 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.276465 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.276489 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:32Z","lastTransitionTime":"2025-09-30T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.306531 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:32 crc kubenswrapper[4810]: E0930 08:03:32.306706 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.309469 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-pnxwm"] Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.310115 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:32 crc kubenswrapper[4810]: E0930 08:03:32.310198 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.333180 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.354016 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.372797 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.380431 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.380479 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.380501 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.380529 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.380546 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:32Z","lastTransitionTime":"2025-09-30T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.395770 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.411030 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.411157 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wdnx\" (UniqueName: \"kubernetes.io/projected/136360fc-f06a-4b28-bbe6-b8cefcac4fda-kube-api-access-4wdnx\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.418820 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.444459 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.477390 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 08:03:28.127246 6117 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:28.127302 6117 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:28.127307 6117 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:28.127317 6117 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:28.127335 6117 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 08:03:28.127346 6117 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 08:03:28.127351 6117 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 08:03:28.127364 6117 factory.go:656] Stopping watch factory\\\\nI0930 08:03:28.127374 6117 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:28.127367 6117 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 08:03:28.127387 6117 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:28.127410 6117 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 08:03:28.127409 6117 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:28.127416 6117 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 08:03:28.127425 6117 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:2910\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:29Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.590716 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.591154 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:29.591222 6235 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 08:03:29.591236 6235 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 08:03:29.591256 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:29.591310 6235 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 08:03:29.591335 6235 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:29.591348 6235 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 08:03:29.591366 6235 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:29.591379 6235 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:29.591427 6235 factory.go:656] Stopping watch factory\\\\nI0930 08:03:29.591442 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:29.591451 6235 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:29.591466 6235 handler.go:208] Removed *v1.Node event handler 2\\\\nI09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.483413 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.483467 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.483487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.483514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.483562 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:32Z","lastTransitionTime":"2025-09-30T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.495603 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.510646 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.512075 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.512162 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wdnx\" (UniqueName: \"kubernetes.io/projected/136360fc-f06a-4b28-bbe6-b8cefcac4fda-kube-api-access-4wdnx\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.512302 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:32 crc kubenswrapper[4810]: E0930 08:03:32.512462 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:32 crc kubenswrapper[4810]: E0930 08:03:32.512540 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs podName:136360fc-f06a-4b28-bbe6-b8cefcac4fda nodeName:}" failed. No retries permitted until 2025-09-30 08:03:33.012517175 +0000 UTC m=+36.464716482 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs") pod "network-metrics-daemon-pnxwm" (UID: "136360fc-f06a-4b28-bbe6-b8cefcac4fda") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.517992 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/022a6cd0-9f4e-4ee1-a6c5-68568c7e4029-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jlk85\" (UID: \"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.526443 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.546374 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.548219 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wdnx\" (UniqueName: \"kubernetes.io/projected/136360fc-f06a-4b28-bbe6-b8cefcac4fda-kube-api-access-4wdnx\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.564835 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.583249 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.587556 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.587693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.587719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.587762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.587788 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:32Z","lastTransitionTime":"2025-09-30T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.589187 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.600853 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: W0930 08:03:32.612972 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod022a6cd0_9f4e_4ee1_a6c5_68568c7e4029.slice/crio-f7616818a357c223436d15e79e4fbd2032960c1bb8b6a3f7cec66b5a1982da84 WatchSource:0}: Error finding container f7616818a357c223436d15e79e4fbd2032960c1bb8b6a3f7cec66b5a1982da84: Status 404 returned error can't find the container with id f7616818a357c223436d15e79e4fbd2032960c1bb8b6a3f7cec66b5a1982da84 Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.626197 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.642442 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.659906 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:32Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.690685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.690722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.690731 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.690747 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.690773 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:32Z","lastTransitionTime":"2025-09-30T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.794029 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.794074 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.794087 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.794108 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.794121 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:32Z","lastTransitionTime":"2025-09-30T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.895776 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.896044 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.896052 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.896066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.896077 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:32Z","lastTransitionTime":"2025-09-30T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.998637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.998682 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.998696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.998713 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:32 crc kubenswrapper[4810]: I0930 08:03:32.998725 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:32Z","lastTransitionTime":"2025-09-30T08:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.017555 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.017717 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.017786 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs podName:136360fc-f06a-4b28-bbe6-b8cefcac4fda nodeName:}" failed. No retries permitted until 2025-09-30 08:03:34.017771004 +0000 UTC m=+37.469970261 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs") pod "network-metrics-daemon-pnxwm" (UID: "136360fc-f06a-4b28-bbe6-b8cefcac4fda") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.101753 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.101795 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.101807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.101828 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.101842 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.204155 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.204197 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.204208 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.204225 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.204237 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.219027 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.219094 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219123 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:03:49.219103837 +0000 UTC m=+52.671303104 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.219149 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.219195 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219215 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219232 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219254 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219299 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219322 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:49.219311593 +0000 UTC m=+52.671510860 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219340 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:49.219332274 +0000 UTC m=+52.671531541 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219341 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219365 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.219232 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219377 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219522 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:49.219511119 +0000 UTC m=+52.671710386 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219392 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.219560 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:03:49.21955241 +0000 UTC m=+52.671751677 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.305556 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.305635 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.305735 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.305821 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.307400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.307434 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.307445 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.307460 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.307472 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.411144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.411244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.411262 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.411319 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.411337 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.514241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.514358 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.514387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.514421 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.514445 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.539224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.539312 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.539333 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.539359 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.539378 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.561765 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.567996 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.568057 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.568082 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.568109 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.568127 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.590553 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.597960 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.598028 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.598045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.598072 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.598090 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.618889 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" event={"ID":"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029","Type":"ContainerStarted","Data":"f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.618982 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" event={"ID":"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029","Type":"ContainerStarted","Data":"b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.619009 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" event={"ID":"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029","Type":"ContainerStarted","Data":"f7616818a357c223436d15e79e4fbd2032960c1bb8b6a3f7cec66b5a1982da84"} Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.625234 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.630766 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.630835 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.630859 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.630894 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.630919 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.647768 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.656891 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.663593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.663851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.663993 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.664137 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.664260 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.667158 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.685372 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: E0930 08:03:33.685612 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.688000 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.688068 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.688086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.688112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.688130 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.702186 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.719450 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.739482 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.761446 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.782534 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.791770 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.791835 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.791852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.791878 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.791895 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.802606 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.823940 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.845340 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.868930 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.894730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.894784 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.894797 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.894813 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.894822 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.901198 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 08:03:28.127246 6117 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:28.127302 6117 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:28.127307 6117 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:28.127317 6117 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:28.127335 6117 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 08:03:28.127346 6117 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 08:03:28.127351 6117 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 08:03:28.127364 6117 factory.go:656] Stopping watch factory\\\\nI0930 08:03:28.127374 6117 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:28.127367 6117 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 08:03:28.127387 6117 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:28.127410 6117 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 08:03:28.127409 6117 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:28.127416 6117 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 08:03:28.127425 6117 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:2910\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:29Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.590716 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.591154 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:29.591222 6235 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 08:03:29.591236 6235 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 08:03:29.591256 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:29.591310 6235 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 08:03:29.591335 6235 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:29.591348 6235 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 08:03:29.591366 6235 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:29.591379 6235 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:29.591427 6235 factory.go:656] Stopping watch factory\\\\nI0930 08:03:29.591442 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:29.591451 6235 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:29.591466 6235 handler.go:208] Removed *v1.Node event handler 2\\\\nI09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.918123 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.939832 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.958750 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.976971 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.995184 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:33Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.997558 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.997628 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.997647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.997672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:33 crc kubenswrapper[4810]: I0930 08:03:33.997694 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:33Z","lastTransitionTime":"2025-09-30T08:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.026431 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:34 crc kubenswrapper[4810]: E0930 08:03:34.026702 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:34 crc kubenswrapper[4810]: E0930 08:03:34.026813 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs podName:136360fc-f06a-4b28-bbe6-b8cefcac4fda nodeName:}" failed. No retries permitted until 2025-09-30 08:03:36.026783358 +0000 UTC m=+39.478982665 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs") pod "network-metrics-daemon-pnxwm" (UID: "136360fc-f06a-4b28-bbe6-b8cefcac4fda") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.100910 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.100985 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.101003 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.101028 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.101047 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:34Z","lastTransitionTime":"2025-09-30T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.205127 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.205189 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.205207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.205231 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.205251 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:34Z","lastTransitionTime":"2025-09-30T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.305494 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.305611 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:34 crc kubenswrapper[4810]: E0930 08:03:34.305682 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:34 crc kubenswrapper[4810]: E0930 08:03:34.305846 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.306776 4810 scope.go:117] "RemoveContainer" containerID="a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.309058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.309134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.309153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.309182 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.309203 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:34Z","lastTransitionTime":"2025-09-30T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.412582 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.412629 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.412646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.412669 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.412686 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:34Z","lastTransitionTime":"2025-09-30T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.515947 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.515985 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.515995 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.516020 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.516033 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:34Z","lastTransitionTime":"2025-09-30T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.620906 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.620938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.620951 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.620971 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.620982 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:34Z","lastTransitionTime":"2025-09-30T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.624470 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.626803 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d"} Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.627616 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.679396 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.697114 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.709178 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.723669 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.723715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.723728 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.723747 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.723758 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:34Z","lastTransitionTime":"2025-09-30T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.726620 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.743607 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.766136 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 08:03:28.127246 6117 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:28.127302 6117 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:28.127307 6117 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:28.127317 6117 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:28.127335 6117 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 08:03:28.127346 6117 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 08:03:28.127351 6117 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 08:03:28.127364 6117 factory.go:656] Stopping watch factory\\\\nI0930 08:03:28.127374 6117 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:28.127367 6117 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 08:03:28.127387 6117 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:28.127410 6117 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 08:03:28.127409 6117 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:28.127416 6117 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 08:03:28.127425 6117 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:2910\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:29Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.590716 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.591154 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:29.591222 6235 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 08:03:29.591236 6235 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 08:03:29.591256 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:29.591310 6235 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 08:03:29.591335 6235 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:29.591348 6235 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 08:03:29.591366 6235 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:29.591379 6235 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:29.591427 6235 factory.go:656] Stopping watch factory\\\\nI0930 08:03:29.591442 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:29.591451 6235 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:29.591466 6235 handler.go:208] Removed *v1.Node event handler 2\\\\nI09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.782746 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.798338 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.813858 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.825914 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.825943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.825954 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.825970 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.825982 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:34Z","lastTransitionTime":"2025-09-30T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.830687 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.843725 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.860976 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.876163 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.891463 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.904617 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.920601 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.929033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.929074 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.929090 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.929118 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.929140 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:34Z","lastTransitionTime":"2025-09-30T08:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:34 crc kubenswrapper[4810]: I0930 08:03:34.934564 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:34Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.031909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.031980 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.031998 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.032028 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.032047 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:35Z","lastTransitionTime":"2025-09-30T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.136086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.136135 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.136152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.136175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.136194 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:35Z","lastTransitionTime":"2025-09-30T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.239426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.239496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.239522 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.239553 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.239577 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:35Z","lastTransitionTime":"2025-09-30T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.305908 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.306048 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:35 crc kubenswrapper[4810]: E0930 08:03:35.306114 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:35 crc kubenswrapper[4810]: E0930 08:03:35.306235 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.342399 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.342433 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.342441 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.342454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.342463 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:35Z","lastTransitionTime":"2025-09-30T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.445635 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.445667 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.445678 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.445694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.445703 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:35Z","lastTransitionTime":"2025-09-30T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.549065 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.549120 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.549133 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.549153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.549166 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:35Z","lastTransitionTime":"2025-09-30T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.652240 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.652322 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.652338 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.652364 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.652382 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:35Z","lastTransitionTime":"2025-09-30T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.755111 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.755170 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.755187 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.755212 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.755229 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:35Z","lastTransitionTime":"2025-09-30T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.859450 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.859576 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.859598 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.859631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.859650 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:35Z","lastTransitionTime":"2025-09-30T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.963062 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.963154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.963202 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.963227 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:35 crc kubenswrapper[4810]: I0930 08:03:35.963253 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:35Z","lastTransitionTime":"2025-09-30T08:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.046956 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:36 crc kubenswrapper[4810]: E0930 08:03:36.047221 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:36 crc kubenswrapper[4810]: E0930 08:03:36.047365 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs podName:136360fc-f06a-4b28-bbe6-b8cefcac4fda nodeName:}" failed. No retries permitted until 2025-09-30 08:03:40.047331688 +0000 UTC m=+43.499530995 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs") pod "network-metrics-daemon-pnxwm" (UID: "136360fc-f06a-4b28-bbe6-b8cefcac4fda") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.067219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.067310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.067334 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.067361 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.067380 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:36Z","lastTransitionTime":"2025-09-30T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.170657 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.170723 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.170741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.170766 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.170787 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:36Z","lastTransitionTime":"2025-09-30T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.273989 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.274076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.274099 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.274134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.274164 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:36Z","lastTransitionTime":"2025-09-30T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.305811 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.305830 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:36 crc kubenswrapper[4810]: E0930 08:03:36.306218 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:36 crc kubenswrapper[4810]: E0930 08:03:36.306398 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.382387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.382458 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.382505 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.382540 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.382561 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:36Z","lastTransitionTime":"2025-09-30T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.486047 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.486106 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.486122 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.486163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.486182 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:36Z","lastTransitionTime":"2025-09-30T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.590355 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.590436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.590454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.590480 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.590498 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:36Z","lastTransitionTime":"2025-09-30T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.693774 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.693832 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.693849 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.693874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.693892 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:36Z","lastTransitionTime":"2025-09-30T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.796817 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.796875 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.796891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.796916 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.796933 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:36Z","lastTransitionTime":"2025-09-30T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.899707 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.899786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.899811 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.899840 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:36 crc kubenswrapper[4810]: I0930 08:03:36.899861 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:36Z","lastTransitionTime":"2025-09-30T08:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.002673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.002730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.002749 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.002774 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.002790 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:37Z","lastTransitionTime":"2025-09-30T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.105464 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.105516 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.105533 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.105558 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.105575 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:37Z","lastTransitionTime":"2025-09-30T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.207925 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.208076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.208100 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.208130 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.208153 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:37Z","lastTransitionTime":"2025-09-30T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.305838 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:37 crc kubenswrapper[4810]: E0930 08:03:37.306023 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.305866 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:37 crc kubenswrapper[4810]: E0930 08:03:37.306247 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.311887 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.312748 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.312807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.312843 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.312862 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:37Z","lastTransitionTime":"2025-09-30T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.325940 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.360262 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.375994 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.393302 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.415437 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.415497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.415515 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.415542 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.415560 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:37Z","lastTransitionTime":"2025-09-30T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.420907 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.444298 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.469008 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.503258 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1875e88bcc6bb683fb6bb880c3ca71f9a1df3b1630a20cc276aef89561f06f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:28Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 08:03:28.127246 6117 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:28.127302 6117 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:28.127307 6117 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:28.127317 6117 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:28.127335 6117 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0930 08:03:28.127346 6117 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0930 08:03:28.127351 6117 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0930 08:03:28.127364 6117 factory.go:656] Stopping watch factory\\\\nI0930 08:03:28.127374 6117 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:28.127367 6117 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 08:03:28.127387 6117 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:28.127410 6117 handler.go:208] Removed *v1.Node event handler 7\\\\nI0930 08:03:28.127409 6117 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:28.127416 6117 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 08:03:28.127425 6117 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:2910\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:29Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.590716 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.591154 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:29.591222 6235 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 08:03:29.591236 6235 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 08:03:29.591256 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:29.591310 6235 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 08:03:29.591335 6235 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:29.591348 6235 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 08:03:29.591366 6235 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:29.591379 6235 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:29.591427 6235 factory.go:656] Stopping watch factory\\\\nI0930 08:03:29.591442 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:29.591451 6235 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:29.591466 6235 handler.go:208] Removed *v1.Node event handler 2\\\\nI09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.519763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.525459 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.525535 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.525573 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.525610 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:37Z","lastTransitionTime":"2025-09-30T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.529990 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.578331 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.595568 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.616754 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.628191 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.628229 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.628244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.628292 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.628311 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:37Z","lastTransitionTime":"2025-09-30T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.635056 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.653972 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.674970 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.690634 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.710100 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:37Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.731363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.731422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.731441 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.731468 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.731487 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:37Z","lastTransitionTime":"2025-09-30T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.835019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.835096 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.835115 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.835140 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.835258 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:37Z","lastTransitionTime":"2025-09-30T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.938240 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.938350 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.938369 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.938399 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:37 crc kubenswrapper[4810]: I0930 08:03:37.938418 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:37Z","lastTransitionTime":"2025-09-30T08:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.041750 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.041835 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.041855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.041885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.041908 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:38Z","lastTransitionTime":"2025-09-30T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.145672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.145735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.145752 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.145778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.145796 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:38Z","lastTransitionTime":"2025-09-30T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.249623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.249679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.249696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.249722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.249740 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:38Z","lastTransitionTime":"2025-09-30T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.306086 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:38 crc kubenswrapper[4810]: E0930 08:03:38.306330 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.306880 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:38 crc kubenswrapper[4810]: E0930 08:03:38.306988 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.353471 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.353528 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.353545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.353570 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.353588 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:38Z","lastTransitionTime":"2025-09-30T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.457384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.457447 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.457464 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.457491 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.457510 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:38Z","lastTransitionTime":"2025-09-30T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.560562 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.560651 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.560678 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.560719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.560758 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:38Z","lastTransitionTime":"2025-09-30T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.664598 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.664656 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.664676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.664705 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.664724 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:38Z","lastTransitionTime":"2025-09-30T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.768833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.768893 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.768910 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.768933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.768951 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:38Z","lastTransitionTime":"2025-09-30T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.872410 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.872526 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.872549 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.872574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.872594 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:38Z","lastTransitionTime":"2025-09-30T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.976661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.976741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.976764 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.976796 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:38 crc kubenswrapper[4810]: I0930 08:03:38.976819 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:38Z","lastTransitionTime":"2025-09-30T08:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.080813 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.080883 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.080902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.081111 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.081129 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:39Z","lastTransitionTime":"2025-09-30T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.184887 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.184941 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.184959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.184984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.185006 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:39Z","lastTransitionTime":"2025-09-30T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.293912 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.294001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.294025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.294062 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.294087 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:39Z","lastTransitionTime":"2025-09-30T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.305834 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:39 crc kubenswrapper[4810]: E0930 08:03:39.306019 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.306164 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:39 crc kubenswrapper[4810]: E0930 08:03:39.306442 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.400702 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.400793 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.400820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.400850 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.400868 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:39Z","lastTransitionTime":"2025-09-30T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.503299 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.503381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.503401 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.503426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.503442 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:39Z","lastTransitionTime":"2025-09-30T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.606488 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.606567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.606590 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.606624 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.606647 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:39Z","lastTransitionTime":"2025-09-30T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.710320 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.710385 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.710397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.710418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.710431 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:39Z","lastTransitionTime":"2025-09-30T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.812854 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.812926 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.812951 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.812980 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.813003 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:39Z","lastTransitionTime":"2025-09-30T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.916061 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.916132 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.916155 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.916184 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:39 crc kubenswrapper[4810]: I0930 08:03:39.916204 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:39Z","lastTransitionTime":"2025-09-30T08:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.019691 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.019754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.019775 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.019799 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.019819 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:40Z","lastTransitionTime":"2025-09-30T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.109163 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:40 crc kubenswrapper[4810]: E0930 08:03:40.109443 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:40 crc kubenswrapper[4810]: E0930 08:03:40.109542 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs podName:136360fc-f06a-4b28-bbe6-b8cefcac4fda nodeName:}" failed. No retries permitted until 2025-09-30 08:03:48.109519091 +0000 UTC m=+51.561718388 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs") pod "network-metrics-daemon-pnxwm" (UID: "136360fc-f06a-4b28-bbe6-b8cefcac4fda") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.123445 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.123520 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.123545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.123576 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.123599 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:40Z","lastTransitionTime":"2025-09-30T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.226596 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.226653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.226673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.226695 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.226713 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:40Z","lastTransitionTime":"2025-09-30T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.305914 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.305945 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:40 crc kubenswrapper[4810]: E0930 08:03:40.306180 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:40 crc kubenswrapper[4810]: E0930 08:03:40.306253 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.330135 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.330195 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.330219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.330246 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.330263 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:40Z","lastTransitionTime":"2025-09-30T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.433665 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.433740 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.433763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.433793 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.433816 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:40Z","lastTransitionTime":"2025-09-30T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.537177 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.537245 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.537310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.537346 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.537365 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:40Z","lastTransitionTime":"2025-09-30T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.640500 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.640589 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.640613 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.640637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.640654 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:40Z","lastTransitionTime":"2025-09-30T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.743496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.743556 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.743575 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.743601 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.743619 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:40Z","lastTransitionTime":"2025-09-30T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.846402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.846461 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.846484 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.846508 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.846528 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:40Z","lastTransitionTime":"2025-09-30T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.949130 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.949182 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.949199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.949221 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:40 crc kubenswrapper[4810]: I0930 08:03:40.949242 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:40Z","lastTransitionTime":"2025-09-30T08:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.052651 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.052720 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.052738 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.052765 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.052783 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:41Z","lastTransitionTime":"2025-09-30T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.155800 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.155880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.155905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.155939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.155964 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:41Z","lastTransitionTime":"2025-09-30T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.259355 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.259430 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.259457 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.259493 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.259513 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:41Z","lastTransitionTime":"2025-09-30T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.305786 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.305815 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:41 crc kubenswrapper[4810]: E0930 08:03:41.305996 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:41 crc kubenswrapper[4810]: E0930 08:03:41.306055 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.362203 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.362301 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.362329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.362361 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.362384 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:41Z","lastTransitionTime":"2025-09-30T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.464984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.465036 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.465053 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.465077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.465092 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:41Z","lastTransitionTime":"2025-09-30T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.568496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.568697 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.568729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.568834 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.568865 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:41Z","lastTransitionTime":"2025-09-30T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.672303 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.672371 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.672389 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.672416 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.672435 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:41Z","lastTransitionTime":"2025-09-30T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.775196 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.775305 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.775328 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.775354 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.775372 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:41Z","lastTransitionTime":"2025-09-30T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.878932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.879057 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.879073 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.879098 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.879119 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:41Z","lastTransitionTime":"2025-09-30T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.982363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.982743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.982760 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.982786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:41 crc kubenswrapper[4810]: I0930 08:03:41.982803 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:41Z","lastTransitionTime":"2025-09-30T08:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.085373 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.085689 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.085873 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.086134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.086376 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:42Z","lastTransitionTime":"2025-09-30T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.189824 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.189883 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.189902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.190251 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.190340 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:42Z","lastTransitionTime":"2025-09-30T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.293511 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.293678 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.293710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.293742 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.293768 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:42Z","lastTransitionTime":"2025-09-30T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.305883 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.305883 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:42 crc kubenswrapper[4810]: E0930 08:03:42.306053 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:42 crc kubenswrapper[4810]: E0930 08:03:42.306190 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.396528 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.396597 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.396617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.396647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.396666 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:42Z","lastTransitionTime":"2025-09-30T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.500293 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.500349 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.500371 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.500400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.500421 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:42Z","lastTransitionTime":"2025-09-30T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.603409 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.603459 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.603471 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.603488 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.603500 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:42Z","lastTransitionTime":"2025-09-30T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.706094 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.706245 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.706309 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.706361 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.706386 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:42Z","lastTransitionTime":"2025-09-30T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.809009 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.809067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.809083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.809107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.809128 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:42Z","lastTransitionTime":"2025-09-30T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.912629 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.912677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.912693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.912717 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:42 crc kubenswrapper[4810]: I0930 08:03:42.912733 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:42Z","lastTransitionTime":"2025-09-30T08:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.015487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.015543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.015571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.015604 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.015628 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.117715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.117761 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.117777 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.117800 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.117817 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.221032 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.221084 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.221103 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.221353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.221376 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.306244 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.306493 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:43 crc kubenswrapper[4810]: E0930 08:03:43.306649 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:43 crc kubenswrapper[4810]: E0930 08:03:43.306899 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.324515 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.324600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.324627 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.324655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.324674 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.427370 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.427452 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.427471 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.427497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.427516 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.531041 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.531909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.532086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.532238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.532426 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.640557 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.640631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.640653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.640680 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.640703 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.744945 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.745030 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.745049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.745090 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.745116 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.848244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.848345 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.848370 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.848443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.848468 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.887661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.887743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.887763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.887792 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.887811 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: E0930 08:03:43.913917 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:43Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.919029 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.919209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.919238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.919342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.919415 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: E0930 08:03:43.943646 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:43Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.949744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.949863 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.949933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.949970 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.950048 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:43 crc kubenswrapper[4810]: E0930 08:03:43.973407 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:43Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.979260 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.979361 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.979380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.979439 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:43 crc kubenswrapper[4810]: I0930 08:03:43.979461 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:43Z","lastTransitionTime":"2025-09-30T08:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: E0930 08:03:43.999774 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:43Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.004833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.004905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.004922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.004954 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.004975 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:44Z","lastTransitionTime":"2025-09-30T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: E0930 08:03:44.026698 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: E0930 08:03:44.026920 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.029223 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.029417 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.029439 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.029461 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.029477 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:44Z","lastTransitionTime":"2025-09-30T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.132382 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.132485 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.132504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.132529 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.132547 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:44Z","lastTransitionTime":"2025-09-30T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.236107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.236172 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.236194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.236228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.236258 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:44Z","lastTransitionTime":"2025-09-30T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.306487 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.306677 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:44 crc kubenswrapper[4810]: E0930 08:03:44.306754 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:44 crc kubenswrapper[4810]: E0930 08:03:44.306928 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.308051 4810 scope.go:117] "RemoveContainer" containerID="ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.340018 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.340086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.340104 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.340134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.340155 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:44Z","lastTransitionTime":"2025-09-30T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.347478 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.369155 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.389009 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.415939 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.443832 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.443933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.443956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.443988 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.444010 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:44Z","lastTransitionTime":"2025-09-30T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.449310 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:29Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.590716 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.591154 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:29.591222 6235 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 08:03:29.591236 6235 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 08:03:29.591256 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:29.591310 6235 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 08:03:29.591335 6235 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:29.591348 6235 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 08:03:29.591366 6235 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:29.591379 6235 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:29.591427 6235 factory.go:656] Stopping watch factory\\\\nI0930 08:03:29.591442 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:29.591451 6235 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:29.591466 6235 handler.go:208] Removed *v1.Node event handler 2\\\\nI09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.475024 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.497773 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.515019 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.543154 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.548791 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.548842 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.548855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.548877 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.548891 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:44Z","lastTransitionTime":"2025-09-30T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.570234 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.588543 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.611427 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.627755 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.642954 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.652242 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.652352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.652372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.652400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.652422 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:44Z","lastTransitionTime":"2025-09-30T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.658365 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.669262 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/1.log" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.673558 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186"} Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.673756 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.678607 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.697234 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.718436 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.734692 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.756009 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.756170 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.756195 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.756221 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.756240 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:44Z","lastTransitionTime":"2025-09-30T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.762325 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.779689 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.800560 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.815886 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.831877 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.851332 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.859648 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.859735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.859801 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.859831 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.859848 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:44Z","lastTransitionTime":"2025-09-30T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.874970 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.907114 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:29Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.590716 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.591154 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:29.591222 6235 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 08:03:29.591236 6235 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 08:03:29.591256 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:29.591310 6235 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 08:03:29.591335 6235 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:29.591348 6235 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 08:03:29.591366 6235 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:29.591379 6235 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:29.591427 6235 factory.go:656] Stopping watch factory\\\\nI0930 08:03:29.591442 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:29.591451 6235 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:29.591466 6235 handler.go:208] Removed *v1.Node event handler 2\\\\nI09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.923717 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.937028 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.949113 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.962869 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.962935 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.962952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.962978 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.962996 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:44Z","lastTransitionTime":"2025-09-30T08:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.966798 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:44 crc kubenswrapper[4810]: I0930 08:03:44.988345 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:44.999936 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:44Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.009736 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.065821 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.065870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.065886 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.065904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.065917 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:45Z","lastTransitionTime":"2025-09-30T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.168018 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.168112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.168134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.168160 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.168179 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:45Z","lastTransitionTime":"2025-09-30T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.271317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.271396 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.271419 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.271451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.271474 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:45Z","lastTransitionTime":"2025-09-30T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.306247 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.306241 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:45 crc kubenswrapper[4810]: E0930 08:03:45.306540 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:45 crc kubenswrapper[4810]: E0930 08:03:45.306617 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.374741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.374830 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.374852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.374881 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.374902 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:45Z","lastTransitionTime":"2025-09-30T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.478425 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.478511 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.478535 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.478567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.478588 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:45Z","lastTransitionTime":"2025-09-30T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.581958 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.582048 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.582073 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.582113 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.582142 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:45Z","lastTransitionTime":"2025-09-30T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.682840 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/2.log" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.684124 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/1.log" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.685489 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.685688 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.685819 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.685972 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.686158 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:45Z","lastTransitionTime":"2025-09-30T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.689604 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186" exitCode=1 Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.689675 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186"} Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.689749 4810 scope.go:117] "RemoveContainer" containerID="ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.691511 4810 scope.go:117] "RemoveContainer" containerID="c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186" Sep 30 08:03:45 crc kubenswrapper[4810]: E0930 08:03:45.694436 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.716044 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.736551 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.761648 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.781950 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.789613 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.789658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.789677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.789704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.789724 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:45Z","lastTransitionTime":"2025-09-30T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.805617 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.823620 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.861471 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.882814 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.892827 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.892897 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.892919 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.892945 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.892963 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:45Z","lastTransitionTime":"2025-09-30T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.902398 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.924656 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.949142 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.982213 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:29Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.590716 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.591154 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:29.591222 6235 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 08:03:29.591236 6235 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 08:03:29.591256 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:29.591310 6235 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 08:03:29.591335 6235 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:29.591348 6235 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 08:03:29.591366 6235 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:29.591379 6235 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:29.591427 6235 factory.go:656] Stopping watch factory\\\\nI0930 08:03:29.591442 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:29.591451 6235 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:29.591466 6235 handler.go:208] Removed *v1.Node event handler 2\\\\nI09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:45Z\\\",\\\"message\\\":\\\" shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z]\\\\nI0930 08:03:45.398176 6475 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0930 08:03:45.398175 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398186 6475 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0930 08:03:45.398189 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-xrbr6\\\\nI0930 08:03:45.398196 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398206 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-pnxwm\\\\nI0930 08:03:45.398214 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metric\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.997210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.997316 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.997343 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.997369 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:45 crc kubenswrapper[4810]: I0930 08:03:45.997396 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:45Z","lastTransitionTime":"2025-09-30T08:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.007094 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.030084 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.050098 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.072703 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.092154 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.100699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.100763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.100781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.100806 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.100834 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:46Z","lastTransitionTime":"2025-09-30T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.203969 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.204048 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.204069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.204095 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.204113 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:46Z","lastTransitionTime":"2025-09-30T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.305405 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.305433 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:46 crc kubenswrapper[4810]: E0930 08:03:46.305577 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:46 crc kubenswrapper[4810]: E0930 08:03:46.305743 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.307584 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.307648 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.307671 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.307698 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.307718 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:46Z","lastTransitionTime":"2025-09-30T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.410952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.411038 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.411061 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.411094 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.411111 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:46Z","lastTransitionTime":"2025-09-30T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.515853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.515920 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.515938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.515965 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.515985 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:46Z","lastTransitionTime":"2025-09-30T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.618923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.618963 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.618980 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.619003 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.619020 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:46Z","lastTransitionTime":"2025-09-30T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.696194 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/2.log" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.721743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.721779 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.721796 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.721816 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.721832 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:46Z","lastTransitionTime":"2025-09-30T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.804497 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.825010 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.826047 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.826313 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.826518 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.826743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.827169 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:46Z","lastTransitionTime":"2025-09-30T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.846822 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.868300 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.891508 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.922738 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:29Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.590716 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.591154 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:29.591222 6235 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 08:03:29.591236 6235 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 08:03:29.591256 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:29.591310 6235 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 08:03:29.591335 6235 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:29.591348 6235 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 08:03:29.591366 6235 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:29.591379 6235 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:29.591427 6235 factory.go:656] Stopping watch factory\\\\nI0930 08:03:29.591442 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:29.591451 6235 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:29.591466 6235 handler.go:208] Removed *v1.Node event handler 2\\\\nI09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:45Z\\\",\\\"message\\\":\\\" shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z]\\\\nI0930 08:03:45.398176 6475 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0930 08:03:45.398175 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398186 6475 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0930 08:03:45.398189 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-xrbr6\\\\nI0930 08:03:45.398196 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398206 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-pnxwm\\\\nI0930 08:03:45.398214 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metric\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.930903 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.930968 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.930985 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.931011 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.931032 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:46Z","lastTransitionTime":"2025-09-30T08:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.946030 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.967323 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:46 crc kubenswrapper[4810]: I0930 08:03:46.984293 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:46Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.002744 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.024471 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.034412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.034466 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.034487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.034517 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.034539 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:47Z","lastTransitionTime":"2025-09-30T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.053902 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.109159 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.126788 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.136651 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.136684 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.136698 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.136716 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.136730 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:47Z","lastTransitionTime":"2025-09-30T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.137834 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.160604 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.176354 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.193993 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.239984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.240039 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.240057 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.240086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.240104 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:47Z","lastTransitionTime":"2025-09-30T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.306177 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.306332 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:47 crc kubenswrapper[4810]: E0930 08:03:47.306428 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:47 crc kubenswrapper[4810]: E0930 08:03:47.306613 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.337784 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae579df43ea1bb6356860f27516b3d4582eaa4a21d44172a6e244343760c87b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:29Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.590716 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 08:03:29.591154 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0930 08:03:29.591222 6235 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0930 08:03:29.591236 6235 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0930 08:03:29.591256 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0930 08:03:29.591310 6235 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0930 08:03:29.591335 6235 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0930 08:03:29.591348 6235 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0930 08:03:29.591366 6235 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 08:03:29.591379 6235 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 08:03:29.591427 6235 factory.go:656] Stopping watch factory\\\\nI0930 08:03:29.591442 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0930 08:03:29.591451 6235 ovnkube.go:599] Stopped ovnkube\\\\nI0930 08:03:29.591466 6235 handler.go:208] Removed *v1.Node event handler 2\\\\nI09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:45Z\\\",\\\"message\\\":\\\" shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z]\\\\nI0930 08:03:45.398176 6475 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0930 08:03:45.398175 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398186 6475 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0930 08:03:45.398189 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-xrbr6\\\\nI0930 08:03:45.398196 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398206 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-pnxwm\\\\nI0930 08:03:45.398214 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metric\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.343061 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.343111 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.343129 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.343155 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.343173 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:47Z","lastTransitionTime":"2025-09-30T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.363564 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.382931 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.406101 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.433852 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.445833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.446120 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.446519 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.446848 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.447011 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:47Z","lastTransitionTime":"2025-09-30T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.456040 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.479668 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.497206 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.522439 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.544043 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.547812 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.549236 4810 scope.go:117] "RemoveContainer" containerID="c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186" Sep 30 08:03:47 crc kubenswrapper[4810]: E0930 08:03:47.549702 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.550492 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.550674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.550809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.550948 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.551080 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:47Z","lastTransitionTime":"2025-09-30T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.561001 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.579636 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.602766 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.618692 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.652421 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.654509 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.654567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.654592 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.654623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.654647 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:47Z","lastTransitionTime":"2025-09-30T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.668412 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.683863 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.708116 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.722531 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.741529 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.757846 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.757897 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.757915 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.757941 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.757958 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:47Z","lastTransitionTime":"2025-09-30T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.776505 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:45Z\\\",\\\"message\\\":\\\" shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z]\\\\nI0930 08:03:45.398176 6475 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0930 08:03:45.398175 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398186 6475 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0930 08:03:45.398189 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-xrbr6\\\\nI0930 08:03:45.398196 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398206 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-pnxwm\\\\nI0930 08:03:45.398214 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metric\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.798701 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.817873 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.837169 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.858953 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.860487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.861076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.861204 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.861362 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.861492 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:47Z","lastTransitionTime":"2025-09-30T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.878487 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.891719 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.901992 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.916922 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.928492 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.942859 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.956445 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.963380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.963409 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.963417 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.963432 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.963444 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:47Z","lastTransitionTime":"2025-09-30T08:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.968103 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:47 crc kubenswrapper[4810]: I0930 08:03:47.977744 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:47Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.066232 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.066332 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.066352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.066378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.066397 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:48Z","lastTransitionTime":"2025-09-30T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.168619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.168735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.168758 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.168783 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.168802 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:48Z","lastTransitionTime":"2025-09-30T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.201510 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:48 crc kubenswrapper[4810]: E0930 08:03:48.201778 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:48 crc kubenswrapper[4810]: E0930 08:03:48.201859 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs podName:136360fc-f06a-4b28-bbe6-b8cefcac4fda nodeName:}" failed. No retries permitted until 2025-09-30 08:04:04.201836048 +0000 UTC m=+67.654035345 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs") pod "network-metrics-daemon-pnxwm" (UID: "136360fc-f06a-4b28-bbe6-b8cefcac4fda") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.271118 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.271171 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.271190 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.271215 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.271232 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:48Z","lastTransitionTime":"2025-09-30T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.305648 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.305690 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:48 crc kubenswrapper[4810]: E0930 08:03:48.305773 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:48 crc kubenswrapper[4810]: E0930 08:03:48.305938 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.374163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.374195 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.374205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.374218 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.374229 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:48Z","lastTransitionTime":"2025-09-30T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.476731 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.476788 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.476807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.476835 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.476852 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:48Z","lastTransitionTime":"2025-09-30T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.579859 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.579927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.579950 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.579981 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.580007 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:48Z","lastTransitionTime":"2025-09-30T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.722474 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.722546 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.722557 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.722576 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.722586 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:48Z","lastTransitionTime":"2025-09-30T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.825617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.825660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.825671 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.825691 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.825703 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:48Z","lastTransitionTime":"2025-09-30T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.928763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.928810 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.928822 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.928841 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:48 crc kubenswrapper[4810]: I0930 08:03:48.928853 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:48Z","lastTransitionTime":"2025-09-30T08:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.031172 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.031243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.031300 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.031327 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.031345 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:49Z","lastTransitionTime":"2025-09-30T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.134807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.134879 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.134902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.134933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.134955 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:49Z","lastTransitionTime":"2025-09-30T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.219684 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.219855 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.219954 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.219958 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:04:21.219914963 +0000 UTC m=+84.672114260 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.220033 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:04:21.220010446 +0000 UTC m=+84.672209743 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.220069 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.220120 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.220193 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.220320 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.220358 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.220371 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.220382 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.220399 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.220422 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.220436 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:04:21.220421778 +0000 UTC m=+84.672621085 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.220439 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.220464 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 08:04:21.220452289 +0000 UTC m=+84.672651586 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.220512 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 08:04:21.22049525 +0000 UTC m=+84.672694557 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.238355 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.238444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.238496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.238523 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.238544 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:49Z","lastTransitionTime":"2025-09-30T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.306032 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.306079 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.306255 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:49 crc kubenswrapper[4810]: E0930 08:03:49.306454 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.342201 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.342250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.342260 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.342296 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.342307 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:49Z","lastTransitionTime":"2025-09-30T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.445508 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.445628 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.445648 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.445673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.445719 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:49Z","lastTransitionTime":"2025-09-30T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.548466 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.548524 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.548543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.548567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.548587 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:49Z","lastTransitionTime":"2025-09-30T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.652037 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.652097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.652114 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.652138 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.652155 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:49Z","lastTransitionTime":"2025-09-30T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.754658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.754722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.754741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.754762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.754777 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:49Z","lastTransitionTime":"2025-09-30T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.857137 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.857194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.857210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.857231 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.857245 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:49Z","lastTransitionTime":"2025-09-30T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.959602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.959656 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.959671 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.959692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:49 crc kubenswrapper[4810]: I0930 08:03:49.959706 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:49Z","lastTransitionTime":"2025-09-30T08:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.062707 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.062747 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.062761 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.062784 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.062798 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:50Z","lastTransitionTime":"2025-09-30T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.165064 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.165115 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.165128 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.165146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.165157 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:50Z","lastTransitionTime":"2025-09-30T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.270058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.270110 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.270122 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.270144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.270159 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:50Z","lastTransitionTime":"2025-09-30T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.306392 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.306415 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:50 crc kubenswrapper[4810]: E0930 08:03:50.306634 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:50 crc kubenswrapper[4810]: E0930 08:03:50.306803 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.378517 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.378599 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.378649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.378673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.378714 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:50Z","lastTransitionTime":"2025-09-30T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.481297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.481348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.481359 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.481376 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.481388 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:50Z","lastTransitionTime":"2025-09-30T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.584233 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.584354 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.584379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.584412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.584439 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:50Z","lastTransitionTime":"2025-09-30T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.606969 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.623760 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.627851 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.650744 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.677546 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.687089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.687247 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.687311 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.687348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.687372 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:50Z","lastTransitionTime":"2025-09-30T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.694251 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.712369 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.727565 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.763004 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.783640 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.790153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.790237 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.790255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.790310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.790329 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:50Z","lastTransitionTime":"2025-09-30T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.802156 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.816700 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.834746 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.852331 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.877512 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:45Z\\\",\\\"message\\\":\\\" shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z]\\\\nI0930 08:03:45.398176 6475 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0930 08:03:45.398175 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398186 6475 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0930 08:03:45.398189 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-xrbr6\\\\nI0930 08:03:45.398196 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398206 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-pnxwm\\\\nI0930 08:03:45.398214 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metric\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.894560 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.894631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.894651 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.894676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.894695 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:50Z","lastTransitionTime":"2025-09-30T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.899479 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.921586 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.941145 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.957201 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:50Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.998199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.998297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.998317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.998341 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:50 crc kubenswrapper[4810]: I0930 08:03:50.998355 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:50Z","lastTransitionTime":"2025-09-30T08:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.101443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.101514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.101533 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.101565 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.101585 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:51Z","lastTransitionTime":"2025-09-30T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.204112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.204154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.204168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.204356 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.204377 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:51Z","lastTransitionTime":"2025-09-30T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.305605 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.305684 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:51 crc kubenswrapper[4810]: E0930 08:03:51.306369 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:51 crc kubenswrapper[4810]: E0930 08:03:51.306488 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.307365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.307388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.307400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.307424 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.307444 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:51Z","lastTransitionTime":"2025-09-30T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.410083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.410136 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.410153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.410177 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.410192 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:51Z","lastTransitionTime":"2025-09-30T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.512771 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.512815 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.512828 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.512850 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.512863 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:51Z","lastTransitionTime":"2025-09-30T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.616769 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.616847 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.616871 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.616902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.616926 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:51Z","lastTransitionTime":"2025-09-30T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.718975 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.719023 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.719035 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.719054 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.719068 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:51Z","lastTransitionTime":"2025-09-30T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.821888 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.821936 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.821947 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.822007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.822021 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:51Z","lastTransitionTime":"2025-09-30T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.925254 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.925323 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.925335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.925355 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:51 crc kubenswrapper[4810]: I0930 08:03:51.925367 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:51Z","lastTransitionTime":"2025-09-30T08:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.027996 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.028066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.028077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.028096 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.028110 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:52Z","lastTransitionTime":"2025-09-30T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.130946 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.130996 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.131012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.131034 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.131050 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:52Z","lastTransitionTime":"2025-09-30T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.233394 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.233438 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.233450 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.233467 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.233479 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:52Z","lastTransitionTime":"2025-09-30T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.305900 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:52 crc kubenswrapper[4810]: E0930 08:03:52.306048 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.305908 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:52 crc kubenswrapper[4810]: E0930 08:03:52.306235 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.336017 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.336066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.336081 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.336100 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.336112 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:52Z","lastTransitionTime":"2025-09-30T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.438691 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.438746 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.438759 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.438776 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.438788 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:52Z","lastTransitionTime":"2025-09-30T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.541392 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.541470 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.541494 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.541519 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.541540 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:52Z","lastTransitionTime":"2025-09-30T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.645067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.645139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.645153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.645176 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.645191 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:52Z","lastTransitionTime":"2025-09-30T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.747834 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.747877 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.747889 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.747907 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.747919 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:52Z","lastTransitionTime":"2025-09-30T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.851307 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.851376 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.851394 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.851421 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.851440 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:52Z","lastTransitionTime":"2025-09-30T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.955193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.955241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.955253 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.955302 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:52 crc kubenswrapper[4810]: I0930 08:03:52.955317 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:52Z","lastTransitionTime":"2025-09-30T08:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.057769 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.057829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.057843 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.057865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.057882 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:53Z","lastTransitionTime":"2025-09-30T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.160564 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.160620 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.160636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.160662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.160676 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:53Z","lastTransitionTime":"2025-09-30T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.263082 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.263135 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.263147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.263171 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.263184 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:53Z","lastTransitionTime":"2025-09-30T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.305926 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.305991 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:53 crc kubenswrapper[4810]: E0930 08:03:53.306077 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:53 crc kubenswrapper[4810]: E0930 08:03:53.306247 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.366049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.366106 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.366121 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.366145 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.366160 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:53Z","lastTransitionTime":"2025-09-30T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.469846 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.469891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.469902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.469920 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.469936 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:53Z","lastTransitionTime":"2025-09-30T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.572816 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.572874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.572886 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.572907 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.572919 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:53Z","lastTransitionTime":"2025-09-30T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.675684 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.675745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.675764 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.675786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.675798 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:53Z","lastTransitionTime":"2025-09-30T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.778647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.778683 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.778692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.778705 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.778714 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:53Z","lastTransitionTime":"2025-09-30T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.881196 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.881239 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.881250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.881281 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.881294 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:53Z","lastTransitionTime":"2025-09-30T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.983623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.983854 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.983868 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.983885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:53 crc kubenswrapper[4810]: I0930 08:03:53.983897 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:53Z","lastTransitionTime":"2025-09-30T08:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.085938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.086012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.086033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.086142 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.086189 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.189001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.189045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.189056 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.189076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.189089 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.267895 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.267961 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.267984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.268013 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.268035 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: E0930 08:03:54.293451 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:54Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.299001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.299067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.299086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.299113 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.299133 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.305555 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.305564 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:54 crc kubenswrapper[4810]: E0930 08:03:54.305745 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:54 crc kubenswrapper[4810]: E0930 08:03:54.305943 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:54 crc kubenswrapper[4810]: E0930 08:03:54.319811 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:54Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.325297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.325363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.325381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.325407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.325425 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: E0930 08:03:54.344113 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:54Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.349662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.349719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.349736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.349760 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.349789 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: E0930 08:03:54.368334 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:54Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.376959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.377013 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.377029 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.377051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.377065 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: E0930 08:03:54.390738 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:54Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:54 crc kubenswrapper[4810]: E0930 08:03:54.390910 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.392863 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.392937 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.392958 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.392984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.393001 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.495372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.495439 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.495461 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.495482 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.495497 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.597725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.597778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.597792 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.597812 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.597824 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.701412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.701479 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.701502 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.701532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.701555 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.805309 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.805365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.805385 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.805410 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.805427 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.908545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.908600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.908614 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.908636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:54 crc kubenswrapper[4810]: I0930 08:03:54.908652 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:54Z","lastTransitionTime":"2025-09-30T08:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.011767 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.011824 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.011843 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.011867 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.011890 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:55Z","lastTransitionTime":"2025-09-30T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.113883 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.113925 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.113939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.113961 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.113977 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:55Z","lastTransitionTime":"2025-09-30T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.217079 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.217155 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.217179 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.217211 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.217238 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:55Z","lastTransitionTime":"2025-09-30T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.306392 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.306474 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:55 crc kubenswrapper[4810]: E0930 08:03:55.306585 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:55 crc kubenswrapper[4810]: E0930 08:03:55.306769 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.320325 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.320368 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.320380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.320404 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.320418 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:55Z","lastTransitionTime":"2025-09-30T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.423870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.423930 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.423942 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.423960 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.423971 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:55Z","lastTransitionTime":"2025-09-30T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.527631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.527694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.527712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.527737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.527759 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:55Z","lastTransitionTime":"2025-09-30T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.630772 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.630851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.630869 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.630897 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.630917 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:55Z","lastTransitionTime":"2025-09-30T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.734328 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.734386 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.734397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.734417 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.734436 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:55Z","lastTransitionTime":"2025-09-30T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.837523 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.837575 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.837594 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.837619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.837637 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:55Z","lastTransitionTime":"2025-09-30T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.940903 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.940968 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.940989 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.941016 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:55 crc kubenswrapper[4810]: I0930 08:03:55.941035 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:55Z","lastTransitionTime":"2025-09-30T08:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.044401 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.044479 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.044499 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.044526 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.044547 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:56Z","lastTransitionTime":"2025-09-30T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.148499 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.148582 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.148613 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.148649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.148672 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:56Z","lastTransitionTime":"2025-09-30T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.252369 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.252438 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.252455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.252482 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.252495 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:56Z","lastTransitionTime":"2025-09-30T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.305844 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.305844 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:56 crc kubenswrapper[4810]: E0930 08:03:56.306107 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:56 crc kubenswrapper[4810]: E0930 08:03:56.306256 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.355672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.355762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.355782 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.355809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.355831 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:56Z","lastTransitionTime":"2025-09-30T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.458864 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.458910 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.458927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.458949 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.458964 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:56Z","lastTransitionTime":"2025-09-30T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.562843 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.562990 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.563007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.563028 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.563043 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:56Z","lastTransitionTime":"2025-09-30T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.666623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.666691 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.666708 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.666732 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.666750 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:56Z","lastTransitionTime":"2025-09-30T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.768608 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.768658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.768673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.768694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.768708 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:56Z","lastTransitionTime":"2025-09-30T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.870925 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.870978 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.870994 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.871015 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.871030 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:56Z","lastTransitionTime":"2025-09-30T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.973863 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.973924 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.973938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.973959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:56 crc kubenswrapper[4810]: I0930 08:03:56.973973 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:56Z","lastTransitionTime":"2025-09-30T08:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.078199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.078254 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.078297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.078353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.078370 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:57Z","lastTransitionTime":"2025-09-30T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.180776 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.180818 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.180833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.180849 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.180862 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:57Z","lastTransitionTime":"2025-09-30T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.283256 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.283330 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.283347 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.283371 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.283385 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:57Z","lastTransitionTime":"2025-09-30T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.305661 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.305953 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:57 crc kubenswrapper[4810]: E0930 08:03:57.306461 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:57 crc kubenswrapper[4810]: E0930 08:03:57.306722 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.329211 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.351599 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.375014 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.387095 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.387161 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.387188 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.387222 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.387248 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:57Z","lastTransitionTime":"2025-09-30T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.406755 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:45Z\\\",\\\"message\\\":\\\" shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z]\\\\nI0930 08:03:45.398176 6475 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0930 08:03:45.398175 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398186 6475 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0930 08:03:45.398189 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-xrbr6\\\\nI0930 08:03:45.398196 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398206 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-pnxwm\\\\nI0930 08:03:45.398214 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metric\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.429032 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.449576 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.465607 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"197c8d8b-140c-4f9b-8a1d-fb4d3fac7c39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9075278876c411f05f90291e052877301a61910dc63fecad4b45c1991262da79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a46e68eef88de27f5f93c67b1a8545fba574725b5bd10afd09688ef453dd66a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b91b227ecf3dbfe23847ac4649941e0d424a4e8e63368b1b5a2c2598d9ac1894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.485826 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.490777 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.490859 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.490880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.490910 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.490934 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:57Z","lastTransitionTime":"2025-09-30T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.500808 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.521117 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.540634 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.554613 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.573054 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.593635 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.595257 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.595360 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.595387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.595421 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.595447 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:57Z","lastTransitionTime":"2025-09-30T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.612296 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.648991 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.669482 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.688168 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:57Z is after 2025-08-24T17:21:41Z" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.699102 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.699211 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.699243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.699308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.699342 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:57Z","lastTransitionTime":"2025-09-30T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.802935 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.802985 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.803000 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.803022 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.803038 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:57Z","lastTransitionTime":"2025-09-30T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.904884 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.904919 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.904928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.904942 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:57 crc kubenswrapper[4810]: I0930 08:03:57.904951 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:57Z","lastTransitionTime":"2025-09-30T08:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.007357 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.007408 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.007419 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.007437 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.007450 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:58Z","lastTransitionTime":"2025-09-30T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.109515 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.109550 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.109561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.109577 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.109589 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:58Z","lastTransitionTime":"2025-09-30T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.211860 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.211894 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.211905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.211923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.211934 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:58Z","lastTransitionTime":"2025-09-30T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.305885 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:03:58 crc kubenswrapper[4810]: E0930 08:03:58.306026 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.306227 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:03:58 crc kubenswrapper[4810]: E0930 08:03:58.306314 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.314089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.314119 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.314130 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.314147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.314158 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:58Z","lastTransitionTime":"2025-09-30T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.416819 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.416870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.416886 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.416911 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.416928 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:58Z","lastTransitionTime":"2025-09-30T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.519607 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.519676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.519702 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.519734 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.519758 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:58Z","lastTransitionTime":"2025-09-30T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.621935 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.621990 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.622001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.622018 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.622030 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:58Z","lastTransitionTime":"2025-09-30T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.724798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.724851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.724866 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.724888 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.724905 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:58Z","lastTransitionTime":"2025-09-30T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.827552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.827621 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.827637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.827660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.827675 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:58Z","lastTransitionTime":"2025-09-30T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.930179 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.930234 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.930251 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.930292 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:58 crc kubenswrapper[4810]: I0930 08:03:58.930305 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:58Z","lastTransitionTime":"2025-09-30T08:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.033327 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.033379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.033394 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.033415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.033430 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:59Z","lastTransitionTime":"2025-09-30T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.136182 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.136219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.136232 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.136250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.136276 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:59Z","lastTransitionTime":"2025-09-30T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.238291 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.238351 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.238363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.238383 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.238395 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:59Z","lastTransitionTime":"2025-09-30T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.306612 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:03:59 crc kubenswrapper[4810]: E0930 08:03:59.306809 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.306940 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:03:59 crc kubenswrapper[4810]: E0930 08:03:59.307098 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.340918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.340979 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.340999 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.341023 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.341041 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:59Z","lastTransitionTime":"2025-09-30T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.443767 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.443857 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.443879 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.443900 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.443914 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:59Z","lastTransitionTime":"2025-09-30T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.546385 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.546456 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.546475 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.546504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.546524 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:59Z","lastTransitionTime":"2025-09-30T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.650060 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.650136 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.650159 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.650188 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.650211 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:59Z","lastTransitionTime":"2025-09-30T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.753897 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.753996 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.754017 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.754078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.754105 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:59Z","lastTransitionTime":"2025-09-30T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.856877 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.856915 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.856927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.856944 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.856958 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:59Z","lastTransitionTime":"2025-09-30T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.959438 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.959545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.959566 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.959590 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:03:59 crc kubenswrapper[4810]: I0930 08:03:59.959607 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:03:59Z","lastTransitionTime":"2025-09-30T08:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.062094 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.062144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.062157 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.062176 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.062189 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:00Z","lastTransitionTime":"2025-09-30T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.164750 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.164803 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.164815 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.164834 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.164847 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:00Z","lastTransitionTime":"2025-09-30T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.267886 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.267929 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.267965 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.267985 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.268019 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:00Z","lastTransitionTime":"2025-09-30T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.305701 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.305758 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:00 crc kubenswrapper[4810]: E0930 08:04:00.305854 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:00 crc kubenswrapper[4810]: E0930 08:04:00.305934 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.306740 4810 scope.go:117] "RemoveContainer" containerID="c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186" Sep 30 08:04:00 crc kubenswrapper[4810]: E0930 08:04:00.307063 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.370351 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.370402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.370414 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.370433 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.370450 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:00Z","lastTransitionTime":"2025-09-30T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.473944 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.474023 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.474051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.474085 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.474111 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:00Z","lastTransitionTime":"2025-09-30T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.577415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.577488 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.577511 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.577542 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.577568 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:00Z","lastTransitionTime":"2025-09-30T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.680340 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.680388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.680404 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.680427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.680446 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:00Z","lastTransitionTime":"2025-09-30T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.783413 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.783526 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.783552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.783588 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.783612 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:00Z","lastTransitionTime":"2025-09-30T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.886192 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.886225 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.886236 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.886253 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.886284 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:00Z","lastTransitionTime":"2025-09-30T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.988891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.988963 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.988982 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.989008 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:00 crc kubenswrapper[4810]: I0930 08:04:00.989026 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:00Z","lastTransitionTime":"2025-09-30T08:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.091864 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.091936 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.091958 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.091986 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.092005 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:01Z","lastTransitionTime":"2025-09-30T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.195200 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.195248 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.195281 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.195304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.195322 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:01Z","lastTransitionTime":"2025-09-30T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.298559 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.298622 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.298640 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.298665 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.298687 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:01Z","lastTransitionTime":"2025-09-30T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.306078 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:01 crc kubenswrapper[4810]: E0930 08:04:01.306321 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.306576 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:01 crc kubenswrapper[4810]: E0930 08:04:01.306870 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.400919 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.400955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.400966 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.400984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.401022 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:01Z","lastTransitionTime":"2025-09-30T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.503081 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.503112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.503121 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.503134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.503143 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:01Z","lastTransitionTime":"2025-09-30T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.606600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.606643 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.606652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.606669 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.606682 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:01Z","lastTransitionTime":"2025-09-30T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.708585 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.708653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.708671 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.708699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.708719 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:01Z","lastTransitionTime":"2025-09-30T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.811033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.811105 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.811123 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.811148 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.811164 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:01Z","lastTransitionTime":"2025-09-30T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.913779 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.913817 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.913828 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.913844 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:01 crc kubenswrapper[4810]: I0930 08:04:01.913855 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:01Z","lastTransitionTime":"2025-09-30T08:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.021686 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.021755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.021773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.021799 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.021831 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:02Z","lastTransitionTime":"2025-09-30T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.125174 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.125257 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.125313 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.125342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.125365 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:02Z","lastTransitionTime":"2025-09-30T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.237851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.237902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.237913 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.237931 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.237941 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:02Z","lastTransitionTime":"2025-09-30T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.306140 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.306289 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:02 crc kubenswrapper[4810]: E0930 08:04:02.306389 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:02 crc kubenswrapper[4810]: E0930 08:04:02.306534 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.340048 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.340083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.340093 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.340107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.340119 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:02Z","lastTransitionTime":"2025-09-30T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.442298 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.442353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.442365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.442382 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.442396 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:02Z","lastTransitionTime":"2025-09-30T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.545190 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.545277 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.545287 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.545304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.545313 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:02Z","lastTransitionTime":"2025-09-30T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.647984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.648032 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.648048 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.648073 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.648087 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:02Z","lastTransitionTime":"2025-09-30T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.750726 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.750806 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.750828 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.750862 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.750880 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:02Z","lastTransitionTime":"2025-09-30T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.854543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.854599 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.854610 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.854627 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.854638 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:02Z","lastTransitionTime":"2025-09-30T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.957035 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.957103 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.957121 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.957147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:02 crc kubenswrapper[4810]: I0930 08:04:02.957167 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:02Z","lastTransitionTime":"2025-09-30T08:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.059548 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.059593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.059603 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.059623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.059636 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:03Z","lastTransitionTime":"2025-09-30T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.162652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.162713 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.162722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.162742 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.162751 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:03Z","lastTransitionTime":"2025-09-30T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.265640 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.265706 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.265725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.265751 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.265770 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:03Z","lastTransitionTime":"2025-09-30T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.306368 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.306513 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:03 crc kubenswrapper[4810]: E0930 08:04:03.306550 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:03 crc kubenswrapper[4810]: E0930 08:04:03.306792 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.369088 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.369158 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.369176 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.369203 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.369226 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:03Z","lastTransitionTime":"2025-09-30T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.471976 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.472064 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.472085 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.472113 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.472130 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:03Z","lastTransitionTime":"2025-09-30T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.575163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.575233 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.575250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.575285 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.575298 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:03Z","lastTransitionTime":"2025-09-30T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.677551 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.677612 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.677632 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.677657 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.677675 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:03Z","lastTransitionTime":"2025-09-30T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.861509 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.861567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.861585 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.861609 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.861627 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:03Z","lastTransitionTime":"2025-09-30T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.964444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.964497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.964512 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.964534 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:03 crc kubenswrapper[4810]: I0930 08:04:03.964551 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:03Z","lastTransitionTime":"2025-09-30T08:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.067124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.067205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.067225 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.067251 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.067301 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.170228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.170299 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.170311 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.170330 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.170342 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.263832 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:04 crc kubenswrapper[4810]: E0930 08:04:04.264105 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:04:04 crc kubenswrapper[4810]: E0930 08:04:04.264198 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs podName:136360fc-f06a-4b28-bbe6-b8cefcac4fda nodeName:}" failed. No retries permitted until 2025-09-30 08:04:36.264168707 +0000 UTC m=+99.716367974 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs") pod "network-metrics-daemon-pnxwm" (UID: "136360fc-f06a-4b28-bbe6-b8cefcac4fda") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.274201 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.274286 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.274304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.274329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.274348 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.305521 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.305646 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:04 crc kubenswrapper[4810]: E0930 08:04:04.305667 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:04 crc kubenswrapper[4810]: E0930 08:04:04.305869 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.377058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.377107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.377124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.377146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.377166 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.480160 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.480216 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.480235 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.480259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.480298 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.521825 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.521895 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.521918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.521949 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.521971 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: E0930 08:04:04.542677 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:04Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.546631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.546699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.546725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.546760 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.546785 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: E0930 08:04:04.569311 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:04Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.574801 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.574862 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.574874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.574892 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.574904 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: E0930 08:04:04.590532 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:04Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.595605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.595674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.595696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.595727 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.595750 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: E0930 08:04:04.613282 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:04Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.617503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.617553 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.617563 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.617582 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.617592 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: E0930 08:04:04.631991 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:04Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:04 crc kubenswrapper[4810]: E0930 08:04:04.632144 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.633837 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.633866 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.633876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.633892 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.633903 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.736880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.736938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.736958 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.736991 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.737008 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.840261 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.840318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.840326 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.840340 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.840350 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.943348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.943396 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.943411 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.943430 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:04 crc kubenswrapper[4810]: I0930 08:04:04.943443 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:04Z","lastTransitionTime":"2025-09-30T08:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.046105 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.046147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.046157 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.046173 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.046185 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:05Z","lastTransitionTime":"2025-09-30T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.149935 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.149991 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.150006 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.150051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.150066 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:05Z","lastTransitionTime":"2025-09-30T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.253602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.253659 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.253673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.253694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.253706 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:05Z","lastTransitionTime":"2025-09-30T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.306447 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.306567 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:05 crc kubenswrapper[4810]: E0930 08:04:05.306707 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:05 crc kubenswrapper[4810]: E0930 08:04:05.306802 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.356766 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.356825 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.356843 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.356868 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.356885 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:05Z","lastTransitionTime":"2025-09-30T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.463183 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.463252 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.463308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.463336 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.463357 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:05Z","lastTransitionTime":"2025-09-30T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.566329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.566393 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.566410 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.566435 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.566450 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:05Z","lastTransitionTime":"2025-09-30T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.669694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.669768 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.669787 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.669815 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.669835 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:05Z","lastTransitionTime":"2025-09-30T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.773436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.773487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.773503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.773527 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.773544 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:05Z","lastTransitionTime":"2025-09-30T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.872395 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n85f4_f5367714-1633-4695-af4e-4eb6419daa96/kube-multus/0.log" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.872467 4810 generic.go:334] "Generic (PLEG): container finished" podID="f5367714-1633-4695-af4e-4eb6419daa96" containerID="de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab" exitCode=1 Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.872508 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n85f4" event={"ID":"f5367714-1633-4695-af4e-4eb6419daa96","Type":"ContainerDied","Data":"de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.873488 4810 scope.go:117] "RemoveContainer" containerID="de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.877611 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.877669 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.877692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.877722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.877744 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:05Z","lastTransitionTime":"2025-09-30T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.903359 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:05Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.920428 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:05Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.934613 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:05Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.964022 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:45Z\\\",\\\"message\\\":\\\" shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z]\\\\nI0930 08:03:45.398176 6475 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0930 08:03:45.398175 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398186 6475 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0930 08:03:45.398189 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-xrbr6\\\\nI0930 08:03:45.398196 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398206 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-pnxwm\\\\nI0930 08:03:45.398214 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metric\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:05Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.978218 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:05Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.980144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.980207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.980219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.980324 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.980355 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:05Z","lastTransitionTime":"2025-09-30T08:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:05 crc kubenswrapper[4810]: I0930 08:04:05.990380 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:05Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.002763 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"197c8d8b-140c-4f9b-8a1d-fb4d3fac7c39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9075278876c411f05f90291e052877301a61910dc63fecad4b45c1991262da79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a46e68eef88de27f5f93c67b1a8545fba574725b5bd10afd09688ef453dd66a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b91b227ecf3dbfe23847ac4649941e0d424a4e8e63368b1b5a2c2598d9ac1894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.015400 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.029119 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.041863 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:05Z\\\",\\\"message\\\":\\\"2025-09-30T08:03:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292\\\\n2025-09-30T08:03:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292 to /host/opt/cni/bin/\\\\n2025-09-30T08:03:20Z [verbose] multus-daemon started\\\\n2025-09-30T08:03:20Z [verbose] Readiness Indicator file check\\\\n2025-09-30T08:04:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.057074 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.071415 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.082759 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.083015 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.083133 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.083224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.083336 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:06Z","lastTransitionTime":"2025-09-30T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.085816 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.098755 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.110377 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.120819 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.132358 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.142712 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.186097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.186146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.186154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.186169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.186179 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:06Z","lastTransitionTime":"2025-09-30T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.288333 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.288451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.288530 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.288603 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.288669 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:06Z","lastTransitionTime":"2025-09-30T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.305838 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.305920 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:06 crc kubenswrapper[4810]: E0930 08:04:06.305999 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:06 crc kubenswrapper[4810]: E0930 08:04:06.306104 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.391750 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.391864 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.391890 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.391926 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.391952 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:06Z","lastTransitionTime":"2025-09-30T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.494430 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.494480 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.494491 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.494507 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.494519 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:06Z","lastTransitionTime":"2025-09-30T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.596590 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.596650 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.596674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.596705 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.596741 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:06Z","lastTransitionTime":"2025-09-30T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.699057 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.699136 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.699169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.699199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.699220 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:06Z","lastTransitionTime":"2025-09-30T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.802095 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.802149 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.802164 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.802185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.802199 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:06Z","lastTransitionTime":"2025-09-30T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.877038 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n85f4_f5367714-1633-4695-af4e-4eb6419daa96/kube-multus/0.log" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.877120 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n85f4" event={"ID":"f5367714-1633-4695-af4e-4eb6419daa96","Type":"ContainerStarted","Data":"535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857"} Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.903955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.904035 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.904058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.904086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.904104 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:06Z","lastTransitionTime":"2025-09-30T08:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.904744 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.929070 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:45Z\\\",\\\"message\\\":\\\" shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z]\\\\nI0930 08:03:45.398176 6475 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0930 08:03:45.398175 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398186 6475 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0930 08:03:45.398189 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-xrbr6\\\\nI0930 08:03:45.398196 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398206 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-pnxwm\\\\nI0930 08:03:45.398214 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metric\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.948258 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.965167 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.978409 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"197c8d8b-140c-4f9b-8a1d-fb4d3fac7c39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9075278876c411f05f90291e052877301a61910dc63fecad4b45c1991262da79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a46e68eef88de27f5f93c67b1a8545fba574725b5bd10afd09688ef453dd66a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b91b227ecf3dbfe23847ac4649941e0d424a4e8e63368b1b5a2c2598d9ac1894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:06 crc kubenswrapper[4810]: I0930 08:04:06.990426 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:06Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.006561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.006623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.006642 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.006668 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.006686 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:07Z","lastTransitionTime":"2025-09-30T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.007319 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.026671 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:05Z\\\",\\\"message\\\":\\\"2025-09-30T08:03:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292\\\\n2025-09-30T08:03:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292 to /host/opt/cni/bin/\\\\n2025-09-30T08:03:20Z [verbose] multus-daemon started\\\\n2025-09-30T08:03:20Z [verbose] Readiness Indicator file check\\\\n2025-09-30T08:04:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.041634 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.057954 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.095832 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.108973 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.109019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.109033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.109053 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.109068 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:07Z","lastTransitionTime":"2025-09-30T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.121501 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.138224 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.152731 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.165456 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.189771 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.203101 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.212030 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.212072 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.212118 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.212140 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.212165 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:07Z","lastTransitionTime":"2025-09-30T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.216721 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.306561 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:07 crc kubenswrapper[4810]: E0930 08:04:07.306776 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.307089 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:07 crc kubenswrapper[4810]: E0930 08:04:07.307202 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.314198 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.314240 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.314259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.314303 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.314325 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:07Z","lastTransitionTime":"2025-09-30T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.325404 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.344423 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.363659 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"197c8d8b-140c-4f9b-8a1d-fb4d3fac7c39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9075278876c411f05f90291e052877301a61910dc63fecad4b45c1991262da79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a46e68eef88de27f5f93c67b1a8545fba574725b5bd10afd09688ef453dd66a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b91b227ecf3dbfe23847ac4649941e0d424a4e8e63368b1b5a2c2598d9ac1894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.382931 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.413245 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.417839 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.417899 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.417923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.417954 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.417980 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:07Z","lastTransitionTime":"2025-09-30T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.435108 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:05Z\\\",\\\"message\\\":\\\"2025-09-30T08:03:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292\\\\n2025-09-30T08:03:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292 to /host/opt/cni/bin/\\\\n2025-09-30T08:03:20Z [verbose] multus-daemon started\\\\n2025-09-30T08:03:20Z [verbose] Readiness Indicator file check\\\\n2025-09-30T08:04:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.457778 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.478180 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:45Z\\\",\\\"message\\\":\\\" shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z]\\\\nI0930 08:03:45.398176 6475 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0930 08:03:45.398175 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398186 6475 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0930 08:03:45.398189 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-xrbr6\\\\nI0930 08:03:45.398196 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398206 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-pnxwm\\\\nI0930 08:03:45.398214 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metric\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.489406 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.501554 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.514079 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.520370 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.520400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.520411 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.520427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.520439 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:07Z","lastTransitionTime":"2025-09-30T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.528223 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.539879 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.554884 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.563558 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.580239 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.590180 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.601792 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:07Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.622810 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.622876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.622894 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.622920 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.622939 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:07Z","lastTransitionTime":"2025-09-30T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.725180 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.725256 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.725307 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.725336 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.725353 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:07Z","lastTransitionTime":"2025-09-30T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.827587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.827628 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.827642 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.827663 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.827679 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:07Z","lastTransitionTime":"2025-09-30T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.930769 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.930823 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.930836 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.930859 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:07 crc kubenswrapper[4810]: I0930 08:04:07.930871 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:07Z","lastTransitionTime":"2025-09-30T08:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.033239 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.033294 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.033307 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.033325 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.033337 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:08Z","lastTransitionTime":"2025-09-30T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.135633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.135683 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.135694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.135710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.135722 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:08Z","lastTransitionTime":"2025-09-30T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.238690 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.238746 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.238763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.238787 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.238804 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:08Z","lastTransitionTime":"2025-09-30T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.305411 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.305466 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:08 crc kubenswrapper[4810]: E0930 08:04:08.305601 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:08 crc kubenswrapper[4810]: E0930 08:04:08.305738 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.341479 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.341518 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.341528 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.341543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.341551 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:08Z","lastTransitionTime":"2025-09-30T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.444331 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.444377 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.444392 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.444411 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.444424 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:08Z","lastTransitionTime":"2025-09-30T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.546909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.546959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.546975 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.546998 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.547015 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:08Z","lastTransitionTime":"2025-09-30T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.649554 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.649617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.649634 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.649676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.649693 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:08Z","lastTransitionTime":"2025-09-30T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.752526 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.752592 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.752610 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.752641 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.752659 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:08Z","lastTransitionTime":"2025-09-30T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.855708 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.855762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.855781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.855804 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.855821 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:08Z","lastTransitionTime":"2025-09-30T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.958324 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.958387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.958410 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.958440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:08 crc kubenswrapper[4810]: I0930 08:04:08.958465 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:08Z","lastTransitionTime":"2025-09-30T08:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.061742 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.061804 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.061816 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.061831 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.061841 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:09Z","lastTransitionTime":"2025-09-30T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.164515 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.164804 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.164870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.164931 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.165011 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:09Z","lastTransitionTime":"2025-09-30T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.267511 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.267790 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.267872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.267944 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.268019 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:09Z","lastTransitionTime":"2025-09-30T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.305671 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.305674 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:09 crc kubenswrapper[4810]: E0930 08:04:09.305945 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:09 crc kubenswrapper[4810]: E0930 08:04:09.306089 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.370750 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.370787 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.370798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.370829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.370840 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:09Z","lastTransitionTime":"2025-09-30T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.473435 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.473514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.473539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.473574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.473597 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:09Z","lastTransitionTime":"2025-09-30T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.575600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.575862 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.575928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.576005 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.576065 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:09Z","lastTransitionTime":"2025-09-30T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.678706 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.678777 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.678796 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.678823 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.678841 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:09Z","lastTransitionTime":"2025-09-30T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.780928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.781166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.781240 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.781323 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.781388 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:09Z","lastTransitionTime":"2025-09-30T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.884409 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.884637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.884706 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.884778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.884833 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:09Z","lastTransitionTime":"2025-09-30T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.987027 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.987071 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.987082 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.987099 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:09 crc kubenswrapper[4810]: I0930 08:04:09.987111 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:09Z","lastTransitionTime":"2025-09-30T08:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.090089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.090388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.090494 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.090594 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.090679 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:10Z","lastTransitionTime":"2025-09-30T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.193521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.193838 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.194473 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.194547 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.194630 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:10Z","lastTransitionTime":"2025-09-30T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.297506 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.297638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.297717 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.297776 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.297829 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:10Z","lastTransitionTime":"2025-09-30T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.306132 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:10 crc kubenswrapper[4810]: E0930 08:04:10.306286 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.306136 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:10 crc kubenswrapper[4810]: E0930 08:04:10.306470 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.400241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.400300 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.400310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.400325 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.400336 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:10Z","lastTransitionTime":"2025-09-30T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.502440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.502496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.502507 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.502523 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.502533 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:10Z","lastTransitionTime":"2025-09-30T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.605255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.605353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.605372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.605397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.605416 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:10Z","lastTransitionTime":"2025-09-30T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.708117 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.708178 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.708198 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.708223 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.708240 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:10Z","lastTransitionTime":"2025-09-30T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.811668 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.811732 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.811756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.811783 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.811800 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:10Z","lastTransitionTime":"2025-09-30T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.913612 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.913695 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.913709 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.913730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:10 crc kubenswrapper[4810]: I0930 08:04:10.913774 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:10Z","lastTransitionTime":"2025-09-30T08:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.016538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.016598 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.016612 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.016630 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.016643 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:11Z","lastTransitionTime":"2025-09-30T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.119647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.119704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.119718 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.119736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.119749 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:11Z","lastTransitionTime":"2025-09-30T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.223179 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.223233 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.223250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.223299 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.223319 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:11Z","lastTransitionTime":"2025-09-30T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.305695 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:11 crc kubenswrapper[4810]: E0930 08:04:11.305944 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.306056 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:11 crc kubenswrapper[4810]: E0930 08:04:11.310581 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.330891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.330978 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.331001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.331034 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.331058 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:11Z","lastTransitionTime":"2025-09-30T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.434644 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.434726 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.434749 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.434781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.434803 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:11Z","lastTransitionTime":"2025-09-30T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.538047 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.538101 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.538118 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.538142 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.538163 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:11Z","lastTransitionTime":"2025-09-30T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.641413 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.641457 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.641473 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.641497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.641512 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:11Z","lastTransitionTime":"2025-09-30T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.744352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.744450 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.744503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.744529 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.744581 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:11Z","lastTransitionTime":"2025-09-30T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.847443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.847499 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.847516 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.847542 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.847561 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:11Z","lastTransitionTime":"2025-09-30T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.949653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.949720 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.949740 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.949765 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:11 crc kubenswrapper[4810]: I0930 08:04:11.949783 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:11Z","lastTransitionTime":"2025-09-30T08:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.052858 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.052921 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.052938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.052966 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.052987 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:12Z","lastTransitionTime":"2025-09-30T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.156516 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.156605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.156641 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.156674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.156695 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:12Z","lastTransitionTime":"2025-09-30T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.260307 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.260362 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.260374 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.260394 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.260410 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:12Z","lastTransitionTime":"2025-09-30T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.306015 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.306092 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:12 crc kubenswrapper[4810]: E0930 08:04:12.306262 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:12 crc kubenswrapper[4810]: E0930 08:04:12.306423 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.363971 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.364121 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.364919 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.365258 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.365356 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:12Z","lastTransitionTime":"2025-09-30T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.469107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.469180 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.469198 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.469228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.469245 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:12Z","lastTransitionTime":"2025-09-30T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.572132 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.572174 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.572184 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.572202 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.572214 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:12Z","lastTransitionTime":"2025-09-30T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.675439 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.675503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.675513 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.675534 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.675548 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:12Z","lastTransitionTime":"2025-09-30T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.778787 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.778896 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.778921 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.778952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.778970 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:12Z","lastTransitionTime":"2025-09-30T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.882153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.882196 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.882207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.882225 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.882235 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:12Z","lastTransitionTime":"2025-09-30T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.985805 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.985923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.985942 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.985966 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:12 crc kubenswrapper[4810]: I0930 08:04:12.985983 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:12Z","lastTransitionTime":"2025-09-30T08:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.088563 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.088626 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.088637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.088657 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.088672 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:13Z","lastTransitionTime":"2025-09-30T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.191153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.191216 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.191235 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.191260 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.191304 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:13Z","lastTransitionTime":"2025-09-30T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.294912 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.294981 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.295003 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.295037 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.295058 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:13Z","lastTransitionTime":"2025-09-30T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.306455 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.306511 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:13 crc kubenswrapper[4810]: E0930 08:04:13.306808 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:13 crc kubenswrapper[4810]: E0930 08:04:13.306972 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.308041 4810 scope.go:117] "RemoveContainer" containerID="c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.399761 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.400175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.400224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.400250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.400303 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:13Z","lastTransitionTime":"2025-09-30T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.504622 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.504683 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.504703 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.504728 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.504746 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:13Z","lastTransitionTime":"2025-09-30T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.608003 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.608055 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.608071 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.608089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.608103 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:13Z","lastTransitionTime":"2025-09-30T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.718486 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.718550 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.718576 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.718609 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.718633 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:13Z","lastTransitionTime":"2025-09-30T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.821412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.821465 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.821476 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.821494 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.821507 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:13Z","lastTransitionTime":"2025-09-30T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.903775 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/2.log" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.909181 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b"} Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.909909 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.923951 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.924009 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.924022 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.924042 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.924054 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:13Z","lastTransitionTime":"2025-09-30T08:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.925787 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:13Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.949492 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:13Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.963884 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:13Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.977987 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"197c8d8b-140c-4f9b-8a1d-fb4d3fac7c39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9075278876c411f05f90291e052877301a61910dc63fecad4b45c1991262da79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a46e68eef88de27f5f93c67b1a8545fba574725b5bd10afd09688ef453dd66a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b91b227ecf3dbfe23847ac4649941e0d424a4e8e63368b1b5a2c2598d9ac1894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:13Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:13 crc kubenswrapper[4810]: I0930 08:04:13.992870 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:13Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.008985 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.024670 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:05Z\\\",\\\"message\\\":\\\"2025-09-30T08:03:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292\\\\n2025-09-30T08:03:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292 to /host/opt/cni/bin/\\\\n2025-09-30T08:03:20Z [verbose] multus-daemon started\\\\n2025-09-30T08:03:20Z [verbose] Readiness Indicator file check\\\\n2025-09-30T08:04:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.026627 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.026680 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.026692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.026713 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.026725 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.042581 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.068463 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:45Z\\\",\\\"message\\\":\\\" shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z]\\\\nI0930 08:03:45.398176 6475 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0930 08:03:45.398175 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398186 6475 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0930 08:03:45.398189 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-xrbr6\\\\nI0930 08:03:45.398196 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398206 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-pnxwm\\\\nI0930 08:03:45.398214 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metric\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.086315 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.097613 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.109895 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.120788 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.129314 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.129348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.129357 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.129371 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.129381 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.131899 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.146070 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.159861 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.172341 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.190185 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.232532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.232580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.232593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.232611 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.232626 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.305869 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.305938 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:14 crc kubenswrapper[4810]: E0930 08:04:14.306007 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:14 crc kubenswrapper[4810]: E0930 08:04:14.306136 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.336699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.336762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.336779 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.336806 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.336825 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.440567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.440641 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.440664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.440699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.440722 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.544033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.544114 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.544137 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.544168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.544191 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.646003 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.646049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.646064 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.646083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.646097 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.749505 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.749575 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.749601 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.749636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.749663 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.852541 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.852603 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.852619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.852642 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.852658 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.865388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.865426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.865443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.865462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.865475 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: E0930 08:04:14.880191 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.884537 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.884607 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.884631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.884657 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.884676 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: E0930 08:04:14.899794 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.905765 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.905880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.905911 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.905949 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.905977 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.916658 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/3.log" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.917927 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/2.log" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.922199 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b" exitCode=1 Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.922259 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.922377 4810 scope.go:117] "RemoveContainer" containerID="c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.924788 4810 scope.go:117] "RemoveContainer" containerID="08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b" Sep 30 08:04:14 crc kubenswrapper[4810]: E0930 08:04:14.925145 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" Sep 30 08:04:14 crc kubenswrapper[4810]: E0930 08:04:14.928030 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.932827 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.932921 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.933033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.933059 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.933074 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.939240 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: E0930 08:04:14.947782 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.952069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.952101 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.952112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.952129 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.952140 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.952861 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: E0930 08:04:14.964851 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: E0930 08:04:14.965045 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.965797 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.966826 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.966876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.966891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.966915 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.966931 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:14Z","lastTransitionTime":"2025-09-30T08:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:14 crc kubenswrapper[4810]: I0930 08:04:14.989239 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:14Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.003831 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.018112 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"197c8d8b-140c-4f9b-8a1d-fb4d3fac7c39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9075278876c411f05f90291e052877301a61910dc63fecad4b45c1991262da79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a46e68eef88de27f5f93c67b1a8545fba574725b5bd10afd09688ef453dd66a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b91b227ecf3dbfe23847ac4649941e0d424a4e8e63368b1b5a2c2598d9ac1894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.034170 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.050451 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.069385 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:05Z\\\",\\\"message\\\":\\\"2025-09-30T08:03:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292\\\\n2025-09-30T08:03:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292 to /host/opt/cni/bin/\\\\n2025-09-30T08:03:20Z [verbose] multus-daemon started\\\\n2025-09-30T08:03:20Z [verbose] Readiness Indicator file check\\\\n2025-09-30T08:04:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.070531 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.070569 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.070584 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.070602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.070613 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:15Z","lastTransitionTime":"2025-09-30T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.088466 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.115201 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00650dac33cfb1d04f60b2d074c941d23e121ce2f7b543ea99adc15d66d6186\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:03:45Z\\\",\\\"message\\\":\\\" shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:03:45Z is after 2025-08-24T17:21:41Z]\\\\nI0930 08:03:45.398176 6475 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0930 08:03:45.398175 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398186 6475 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0930 08:03:45.398189 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-xrbr6\\\\nI0930 08:03:45.398196 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0930 08:03:45.398206 6475 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-pnxwm\\\\nI0930 08:03:45.398214 6475 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metric\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"3521ms\\\\nI0930 08:04:14.277215 6843 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0930 08:04:14.277217 6843 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-controller-manager/controller-manager]} name:Service_openshift-controller-manager/controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.149:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {cab7c637-a021-4a4d-a4b9-06d63c44316f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0930 08:04:14.277335 6843 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.133558 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.151478 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.162415 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.173294 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.173329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.173339 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.173354 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.173367 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:15Z","lastTransitionTime":"2025-09-30T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.175107 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.192858 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.213038 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.235866 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.276660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.276715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.276725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.276742 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.276753 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:15Z","lastTransitionTime":"2025-09-30T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.306575 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.306667 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:15 crc kubenswrapper[4810]: E0930 08:04:15.306810 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:15 crc kubenswrapper[4810]: E0930 08:04:15.306897 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.379144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.379225 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.379250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.379326 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.379355 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:15Z","lastTransitionTime":"2025-09-30T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.482145 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.482194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.482208 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.482227 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.482240 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:15Z","lastTransitionTime":"2025-09-30T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.585097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.585142 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.585153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.585169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.585179 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:15Z","lastTransitionTime":"2025-09-30T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.687652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.687699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.687711 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.687732 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.687744 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:15Z","lastTransitionTime":"2025-09-30T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.790904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.790955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.790966 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.790984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.790996 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:15Z","lastTransitionTime":"2025-09-30T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.892962 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.893016 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.893029 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.893048 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.893061 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:15Z","lastTransitionTime":"2025-09-30T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.928800 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/3.log" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.933085 4810 scope.go:117] "RemoveContainer" containerID="08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b" Sep 30 08:04:15 crc kubenswrapper[4810]: E0930 08:04:15.933333 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.948957 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.971654 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.982805 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.995642 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"197c8d8b-140c-4f9b-8a1d-fb4d3fac7c39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9075278876c411f05f90291e052877301a61910dc63fecad4b45c1991262da79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a46e68eef88de27f5f93c67b1a8545fba574725b5bd10afd09688ef453dd66a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b91b227ecf3dbfe23847ac4649941e0d424a4e8e63368b1b5a2c2598d9ac1894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:15Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.995646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.996696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.996733 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.996757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:15 crc kubenswrapper[4810]: I0930 08:04:15.996773 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:15Z","lastTransitionTime":"2025-09-30T08:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.012989 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.029592 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.045734 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:05Z\\\",\\\"message\\\":\\\"2025-09-30T08:03:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292\\\\n2025-09-30T08:03:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292 to /host/opt/cni/bin/\\\\n2025-09-30T08:03:20Z [verbose] multus-daemon started\\\\n2025-09-30T08:03:20Z [verbose] Readiness Indicator file check\\\\n2025-09-30T08:04:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.062523 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.081502 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"3521ms\\\\nI0930 08:04:14.277215 6843 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0930 08:04:14.277217 6843 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-controller-manager/controller-manager]} name:Service_openshift-controller-manager/controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.149:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {cab7c637-a021-4a4d-a4b9-06d63c44316f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0930 08:04:14.277335 6843 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:04:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.096705 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.099477 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.099514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.099522 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.099535 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.099544 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:16Z","lastTransitionTime":"2025-09-30T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.111559 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.127604 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.143445 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.154199 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.167575 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.179682 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.190736 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.201984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.202029 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.202040 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.202056 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.202069 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:16Z","lastTransitionTime":"2025-09-30T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.203807 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:16Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.305222 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.305329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.305354 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.305385 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.305406 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.305474 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:16 crc kubenswrapper[4810]: E0930 08:04:16.305566 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.305407 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:16Z","lastTransitionTime":"2025-09-30T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:16 crc kubenswrapper[4810]: E0930 08:04:16.305815 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.408413 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.408500 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.408565 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.408598 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.408621 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:16Z","lastTransitionTime":"2025-09-30T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.511885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.511959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.511978 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.512004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.512023 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:16Z","lastTransitionTime":"2025-09-30T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.614939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.615007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.615090 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.615125 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.615147 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:16Z","lastTransitionTime":"2025-09-30T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.717538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.717587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.717599 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.717620 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.717641 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:16Z","lastTransitionTime":"2025-09-30T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.820012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.820055 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.820067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.820085 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.820099 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:16Z","lastTransitionTime":"2025-09-30T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.923237 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.923346 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.923366 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.923388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:16 crc kubenswrapper[4810]: I0930 08:04:16.923403 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:16Z","lastTransitionTime":"2025-09-30T08:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.027133 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.027197 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.027210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.027229 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.027246 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:17Z","lastTransitionTime":"2025-09-30T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.130996 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.131042 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.131060 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.131087 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.131105 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:17Z","lastTransitionTime":"2025-09-30T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.234635 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.234737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.234769 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.234818 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.234840 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:17Z","lastTransitionTime":"2025-09-30T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.306336 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.306336 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:17 crc kubenswrapper[4810]: E0930 08:04:17.306553 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:17 crc kubenswrapper[4810]: E0930 08:04:17.307136 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.328735 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.340790 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.340869 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.340887 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.340913 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.340959 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:17Z","lastTransitionTime":"2025-09-30T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.350016 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"197c8d8b-140c-4f9b-8a1d-fb4d3fac7c39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9075278876c411f05f90291e052877301a61910dc63fecad4b45c1991262da79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a46e68eef88de27f5f93c67b1a8545fba574725b5bd10afd09688ef453dd66a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b91b227ecf3dbfe23847ac4649941e0d424a4e8e63368b1b5a2c2598d9ac1894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.369634 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.386612 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.410060 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:05Z\\\",\\\"message\\\":\\\"2025-09-30T08:03:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292\\\\n2025-09-30T08:03:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292 to /host/opt/cni/bin/\\\\n2025-09-30T08:03:20Z [verbose] multus-daemon started\\\\n2025-09-30T08:03:20Z [verbose] Readiness Indicator file check\\\\n2025-09-30T08:04:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.436668 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.444431 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.444516 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.444544 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.444578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.444598 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:17Z","lastTransitionTime":"2025-09-30T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.470512 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"3521ms\\\\nI0930 08:04:14.277215 6843 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0930 08:04:14.277217 6843 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-controller-manager/controller-manager]} name:Service_openshift-controller-manager/controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.149:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {cab7c637-a021-4a4d-a4b9-06d63c44316f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0930 08:04:14.277335 6843 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:04:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.489740 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.504778 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.523087 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.540424 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.550403 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.550564 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.550692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.550814 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.550923 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:17Z","lastTransitionTime":"2025-09-30T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.558322 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.580833 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.601259 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.613211 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.626021 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.638780 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.653413 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.653462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.653476 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.653497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.653508 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:17Z","lastTransitionTime":"2025-09-30T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.672933 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:17Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.756073 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.756142 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.756159 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.756595 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.756620 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:17Z","lastTransitionTime":"2025-09-30T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.859565 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.859607 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.859619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.859636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.859647 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:17Z","lastTransitionTime":"2025-09-30T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.967135 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.967175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.967189 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.967215 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:17 crc kubenswrapper[4810]: I0930 08:04:17.967228 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:17Z","lastTransitionTime":"2025-09-30T08:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.070687 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.070732 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.070743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.070757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.070767 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:18Z","lastTransitionTime":"2025-09-30T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.173748 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.173831 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.173855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.173888 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.173910 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:18Z","lastTransitionTime":"2025-09-30T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.277376 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.277425 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.277447 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.277476 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.277499 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:18Z","lastTransitionTime":"2025-09-30T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.305868 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.305878 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:18 crc kubenswrapper[4810]: E0930 08:04:18.306088 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:18 crc kubenswrapper[4810]: E0930 08:04:18.306192 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.380768 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.380837 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.380853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.380880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.380898 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:18Z","lastTransitionTime":"2025-09-30T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.484633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.484725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.484747 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.484771 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.484791 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:18Z","lastTransitionTime":"2025-09-30T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.588187 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.588304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.588333 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.588363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.588383 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:18Z","lastTransitionTime":"2025-09-30T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.691423 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.691532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.691553 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.691617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.691636 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:18Z","lastTransitionTime":"2025-09-30T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.794456 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.794492 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.794500 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.794514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.794524 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:18Z","lastTransitionTime":"2025-09-30T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.897952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.898023 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.898045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.898076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:18 crc kubenswrapper[4810]: I0930 08:04:18.898100 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:18Z","lastTransitionTime":"2025-09-30T08:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.000647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.000704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.000722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.000748 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.000768 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:19Z","lastTransitionTime":"2025-09-30T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.104804 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.104918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.104937 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.104964 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.104980 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:19Z","lastTransitionTime":"2025-09-30T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.208120 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.208483 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.208617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.208801 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.208957 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:19Z","lastTransitionTime":"2025-09-30T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.305699 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:19 crc kubenswrapper[4810]: E0930 08:04:19.305914 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.306021 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:19 crc kubenswrapper[4810]: E0930 08:04:19.306165 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.312504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.312801 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.312986 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.313188 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.313397 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:19Z","lastTransitionTime":"2025-09-30T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.417214 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.417339 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.417359 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.417386 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.417404 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:19Z","lastTransitionTime":"2025-09-30T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.521235 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.521332 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.521352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.521379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.521397 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:19Z","lastTransitionTime":"2025-09-30T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.624382 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.624445 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.624467 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.624495 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.624513 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:19Z","lastTransitionTime":"2025-09-30T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.728110 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.728216 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.728240 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.728304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.728324 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:19Z","lastTransitionTime":"2025-09-30T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.832050 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.832499 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.832668 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.832820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.832962 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:19Z","lastTransitionTime":"2025-09-30T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.936243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.936870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.937100 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.937344 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:19 crc kubenswrapper[4810]: I0930 08:04:19.937540 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:19Z","lastTransitionTime":"2025-09-30T08:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.040711 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.040781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.040806 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.040842 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.040862 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:20Z","lastTransitionTime":"2025-09-30T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.143377 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.143439 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.143456 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.143484 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.143503 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:20Z","lastTransitionTime":"2025-09-30T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.246905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.246969 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.246992 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.247024 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.247048 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:20Z","lastTransitionTime":"2025-09-30T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.306355 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.306388 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:20 crc kubenswrapper[4810]: E0930 08:04:20.306549 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:20 crc kubenswrapper[4810]: E0930 08:04:20.306733 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.350127 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.350660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.350913 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.351117 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.351385 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:20Z","lastTransitionTime":"2025-09-30T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.454361 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.454738 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.454885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.455028 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.455202 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:20Z","lastTransitionTime":"2025-09-30T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.558083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.558130 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.558146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.558166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.558183 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:20Z","lastTransitionTime":"2025-09-30T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.661551 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.661627 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.661646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.661671 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.661688 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:20Z","lastTransitionTime":"2025-09-30T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.765525 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.765601 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.765631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.765664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.765684 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:20Z","lastTransitionTime":"2025-09-30T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.868796 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.869089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.869303 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.869490 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.869656 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:20Z","lastTransitionTime":"2025-09-30T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.972619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.972939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.973068 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.973206 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:20 crc kubenswrapper[4810]: I0930 08:04:20.973413 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:20Z","lastTransitionTime":"2025-09-30T08:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.076729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.076817 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.076841 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.076873 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.076892 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:21Z","lastTransitionTime":"2025-09-30T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.181402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.181468 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.181482 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.181499 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.181511 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:21Z","lastTransitionTime":"2025-09-30T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.249716 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.249985 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.250050 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.250095 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.250144 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250336 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250383 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250386 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250448 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250474 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.250440156 +0000 UTC m=+148.702639463 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250507 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.250493537 +0000 UTC m=+148.702692834 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250474 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250385 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250602 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250607 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.250556779 +0000 UTC m=+148.702756086 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250689 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.250673802 +0000 UTC m=+148.702873109 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250622 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.250755 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.250740344 +0000 UTC m=+148.702939641 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.284983 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.285053 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.285070 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.285097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.285117 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:21Z","lastTransitionTime":"2025-09-30T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.305856 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.306034 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.306606 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:21 crc kubenswrapper[4810]: E0930 08:04:21.306828 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.388976 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.389059 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.389088 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.389122 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.389145 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:21Z","lastTransitionTime":"2025-09-30T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.492082 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.492157 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.492181 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.492214 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.492233 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:21Z","lastTransitionTime":"2025-09-30T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.595191 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.595659 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.595679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.595704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.595722 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:21Z","lastTransitionTime":"2025-09-30T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.699372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.699496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.699513 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.699539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.699559 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:21Z","lastTransitionTime":"2025-09-30T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.802544 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.802593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.802609 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.802632 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.802648 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:21Z","lastTransitionTime":"2025-09-30T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.906189 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.906237 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.906255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.906319 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:21 crc kubenswrapper[4810]: I0930 08:04:21.906343 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:21Z","lastTransitionTime":"2025-09-30T08:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.009734 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.009808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.009827 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.009855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.009876 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:22Z","lastTransitionTime":"2025-09-30T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.113798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.113865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.113885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.113915 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.113934 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:22Z","lastTransitionTime":"2025-09-30T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.217474 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.217539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.217555 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.217584 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.217605 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:22Z","lastTransitionTime":"2025-09-30T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.306380 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:22 crc kubenswrapper[4810]: E0930 08:04:22.306547 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.306406 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:22 crc kubenswrapper[4810]: E0930 08:04:22.306632 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.321331 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.321402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.321419 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.321447 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.321467 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:22Z","lastTransitionTime":"2025-09-30T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.424329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.424397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.424415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.424445 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.424463 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:22Z","lastTransitionTime":"2025-09-30T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.527549 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.527602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.527621 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.527648 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.527667 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:22Z","lastTransitionTime":"2025-09-30T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.630927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.630986 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.631003 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.631027 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.631048 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:22Z","lastTransitionTime":"2025-09-30T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.734598 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.734670 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.734694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.734725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.734744 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:22Z","lastTransitionTime":"2025-09-30T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.838866 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.838956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.838982 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.839019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.839057 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:22Z","lastTransitionTime":"2025-09-30T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.942417 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.942496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.942520 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.942552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:22 crc kubenswrapper[4810]: I0930 08:04:22.942586 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:22Z","lastTransitionTime":"2025-09-30T08:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.045663 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.045750 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.045768 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.045796 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.045817 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:23Z","lastTransitionTime":"2025-09-30T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.148933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.148995 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.149011 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.149038 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.149078 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:23Z","lastTransitionTime":"2025-09-30T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.252576 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.252634 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.252658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.252687 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.252709 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:23Z","lastTransitionTime":"2025-09-30T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.306080 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.306165 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:23 crc kubenswrapper[4810]: E0930 08:04:23.306384 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:23 crc kubenswrapper[4810]: E0930 08:04:23.306550 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.355715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.355790 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.355814 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.355842 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.355865 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:23Z","lastTransitionTime":"2025-09-30T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.458613 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.458725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.458746 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.458771 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.458790 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:23Z","lastTransitionTime":"2025-09-30T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.561162 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.561220 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.561237 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.561263 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.561339 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:23Z","lastTransitionTime":"2025-09-30T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.663683 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.663739 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.663758 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.663785 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.663803 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:23Z","lastTransitionTime":"2025-09-30T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.766411 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.766468 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.766487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.766513 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.766532 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:23Z","lastTransitionTime":"2025-09-30T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.869414 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.869462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.869476 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.869500 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.869515 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:23Z","lastTransitionTime":"2025-09-30T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.972158 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.972510 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.972676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.972853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:23 crc kubenswrapper[4810]: I0930 08:04:23.972985 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:23Z","lastTransitionTime":"2025-09-30T08:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.076631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.076744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.076764 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.076792 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.076809 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:24Z","lastTransitionTime":"2025-09-30T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.179580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.180376 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.180401 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.180426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.180445 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:24Z","lastTransitionTime":"2025-09-30T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.283888 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.283973 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.283993 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.284023 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.284042 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:24Z","lastTransitionTime":"2025-09-30T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.306353 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.306363 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:24 crc kubenswrapper[4810]: E0930 08:04:24.306544 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:24 crc kubenswrapper[4810]: E0930 08:04:24.306715 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.391615 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.391701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.391720 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.391746 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.391768 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:24Z","lastTransitionTime":"2025-09-30T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.495233 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.495329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.495350 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.495379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.495398 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:24Z","lastTransitionTime":"2025-09-30T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.599256 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.599342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.599360 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.599383 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.599401 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:24Z","lastTransitionTime":"2025-09-30T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.702146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.702194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.702211 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.702232 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.702249 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:24Z","lastTransitionTime":"2025-09-30T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.804923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.805025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.805043 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.805067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.805083 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:24Z","lastTransitionTime":"2025-09-30T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.907829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.907914 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.907931 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.907958 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:24 crc kubenswrapper[4810]: I0930 08:04:24.907980 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:24Z","lastTransitionTime":"2025-09-30T08:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.010904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.010971 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.010993 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.011020 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.011039 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.113694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.113761 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.113777 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.113803 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.113819 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.194124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.194939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.195083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.195367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.195536 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: E0930 08:04:25.215835 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.220875 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.220918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.220935 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.220958 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.220974 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: E0930 08:04:25.240923 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.249666 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.249981 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.250164 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.250388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.250642 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: E0930 08:04:25.267132 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.272163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.272399 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.272527 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.272679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.272814 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: E0930 08:04:25.287617 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.292451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.292583 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.292701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.292817 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.292932 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.305830 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.305865 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:25 crc kubenswrapper[4810]: E0930 08:04:25.306489 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:25 crc kubenswrapper[4810]: E0930 08:04:25.306659 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:25 crc kubenswrapper[4810]: E0930 08:04:25.312908 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:25Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:25 crc kubenswrapper[4810]: E0930 08:04:25.313126 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.315204 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.315460 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.315639 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.315841 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.316031 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.419544 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.419623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.419649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.419684 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.419707 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.523794 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.524138 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.524262 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.524400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.524521 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.627456 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.627837 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.628010 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.628219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.628481 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.730516 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.730569 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.730588 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.730611 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.730627 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.833170 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.833495 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.833631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.833812 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.833943 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.937668 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.937734 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.937753 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.937782 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:25 crc kubenswrapper[4810]: I0930 08:04:25.937800 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:25Z","lastTransitionTime":"2025-09-30T08:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.041562 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.041619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.041637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.041662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.041680 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:26Z","lastTransitionTime":"2025-09-30T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.144762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.145178 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.145487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.145737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.145954 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:26Z","lastTransitionTime":"2025-09-30T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.249496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.249569 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.249592 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.249624 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.249646 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:26Z","lastTransitionTime":"2025-09-30T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.306200 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.306210 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:26 crc kubenswrapper[4810]: E0930 08:04:26.306424 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:26 crc kubenswrapper[4810]: E0930 08:04:26.306602 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.353004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.353044 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.353058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.353074 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.353088 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:26Z","lastTransitionTime":"2025-09-30T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.455846 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.455896 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.455911 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.455933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.455948 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:26Z","lastTransitionTime":"2025-09-30T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.559552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.559617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.559634 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.559662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.559680 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:26Z","lastTransitionTime":"2025-09-30T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.663817 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.663887 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.663906 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.663932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.663952 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:26Z","lastTransitionTime":"2025-09-30T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.766642 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.766733 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.766752 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.766778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.766797 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:26Z","lastTransitionTime":"2025-09-30T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.870139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.870308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.870336 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.870363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.870380 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:26Z","lastTransitionTime":"2025-09-30T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.973074 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.973137 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.973161 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.973189 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:26 crc kubenswrapper[4810]: I0930 08:04:26.973210 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:26Z","lastTransitionTime":"2025-09-30T08:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.076466 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.076539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.076562 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.076588 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.076606 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:27Z","lastTransitionTime":"2025-09-30T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.179852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.179926 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.179950 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.179982 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.180008 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:27Z","lastTransitionTime":"2025-09-30T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.283393 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.283463 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.283483 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.283510 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.283529 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:27Z","lastTransitionTime":"2025-09-30T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.306093 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:27 crc kubenswrapper[4810]: E0930 08:04:27.306465 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.306493 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:27 crc kubenswrapper[4810]: E0930 08:04:27.307413 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.307991 4810 scope.go:117] "RemoveContainer" containerID="08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b" Sep 30 08:04:27 crc kubenswrapper[4810]: E0930 08:04:27.309515 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.326095 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.330465 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.348378 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xrbr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5a0024-ccdd-42f8-b7ed-1e395209d05b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://254b719c25cf2354442fafadcc871f6f4039a220e79ef4c046b3a4ed1847cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kvgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xrbr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.381082 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca47f73e-e1bc-4afc-a08c-8d8dff27c1bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb4fa5d7aceec897ef2a7b778a17650e26d678775573e3905b242f80272e9183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22eeb48510a36fce81115b2590496390bac0956f94204a8cb5993cfb5ed8436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10d139659684ec1efdf7d110008b1365266c0dfc27859c6591ea50621d5efcb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c69066b339b68eba94cf7b9bff02d1859763b7293dbab0fbe8e0ade0624538ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc7f881c4d5af6a90dc870aadadbdede7d755e8759580e8a6ce8a3375b6e8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53fb77dbf7a773a4c951ecfd8d170425a9d213f5bdf4a5106c1f7091426680b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6641141b5319015e5f7eae9ba02d3ef458e023fa0cd4c1fab1d1762503aca0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3995945d09db550dd3aea4733c701f77710ff0dc4550ad2ff17664f2de5868\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.387015 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.387068 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.387085 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.387110 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.387127 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:27Z","lastTransitionTime":"2025-09-30T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.398061 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z5md6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"154f9139-cbed-4ce1-a15b-b28a19016fab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5c913b8e7503d842e7a8966e259977d7857228afe971ca69ea814d7d55b21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtts8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z5md6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.418831 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"022a6cd0-9f4e-4ee1-a6c5-68568c7e4029\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8f1c065feace975a6a29d0509380cbfc816b2fecbdb551e8bb6888515cacc6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79be0f988fecb86941aeec1c65c6e7ab8adaad4554ad4a09aaf07c8ca926947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bkb6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jlk85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.442935 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"128f6072-cce3-4aa1-beae-baa0a4f7d03b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bce23734b248461e78a5c57a09f18212e9ce141b96baa63c64cca0250b6cfb2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21cd462311b5241b6b012ca4deea4f4950c994e8ff08763a6ccb56dbd2657c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb5a0edcd52abecc51be421f42bbbd0de8646f4bb1d1dac680716baeb0129b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45987a40d31cb484f5a7277b800c8c8356e82ff67076bfb2a312479a7a903529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a131b4517e4ef5d399c7d8968903551b665820b583613aca98322cd02863445d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12c8781b572a1a789747749b8189e3f2d2302a2c4d14fa51d908c8d94d9f7747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87fe8145ef8171bf854fbca53e1c091203c2cc93b8974ce5c382b1d3024f6891\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wknsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fnzgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.467117 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aacbcdc8-7ef5-473b-8055-145f40040d5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:14Z\\\",\\\"message\\\":\\\"3521ms\\\\nI0930 08:04:14.277215 6843 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0930 08:04:14.277217 6843 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-controller-manager/controller-manager]} name:Service_openshift-controller-manager/controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.149:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {cab7c637-a021-4a4d-a4b9-06d63c44316f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0930 08:04:14.277335 6843 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:04:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx6mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9bkcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.484340 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c3523e-d7ba-4351-8673-719247b4bd57\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f6a7865a76b0430bb22ed405723f7de759241933ddf7c3f639254265879f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5123c84f35e2e9c79e85dece63327d10f975a9a8499f519652c0545ae5f2f9cc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd89bd366551c3a755bd82accfd564e691bae396c5918e93229729a54c9b2032\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b93d59d604fffc21e0e27243658a9d30144a4b766cf034bce1e0e8d418a06d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b47e4258c53ff7976d73a078e0e19b5624d7a29297427cca93a24aea8cbf54\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T08:03:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0930 08:03:13.283666 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 08:03:13.283971 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 08:03:13.285112 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2443619321/tls.crt::/tmp/serving-cert-2443619321/tls.key\\\\\\\"\\\\nI0930 08:03:13.840247 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 08:03:13.843254 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 08:03:13.843285 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 08:03:13.843303 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 08:03:13.843308 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 08:03:13.848711 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0930 08:03:13.848763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848773 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 08:03:13.848783 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 08:03:13.848789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 08:03:13.848796 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 08:03:13.848802 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0930 08:03:13.848841 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0930 08:03:13.850514 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f0cb8a603e266d1f479a0d2f53a7312fd09a1049632e8c62faa3aafecd688d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d22dcdf8325355a8148e9b7eb232df32e76e1f794d60f693555b7e71bc58c091\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.489863 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.489905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.489921 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.489947 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.489962 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:27Z","lastTransitionTime":"2025-09-30T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.501951 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e6c9ba4-ce12-4546-9dbf-651b9441a059\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79e636f711e8e561c038bd4473c05fa54c4a1923299702d88dfeacf4b34c9a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1e61e3e5cc17755dea9e12029a11960ef0c37c5cd259c40143259cb48b2364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d4b68e0486b78235cc3d657a6f308bc525c03551b6293341d21d01d922fa98\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c111a1130f6de21076c69bcf92577003fc4f207a5617d4f427339b6741f52b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.514594 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"197c8d8b-140c-4f9b-8a1d-fb4d3fac7c39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:02:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9075278876c411f05f90291e052877301a61910dc63fecad4b45c1991262da79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a46e68eef88de27f5f93c67b1a8545fba574725b5bd10afd09688ef453dd66a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:02:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b91b227ecf3dbfe23847ac4649941e0d424a4e8e63368b1b5a2c2598d9ac1894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c8ae87cc3d92906acfdc35ca05c1ec393a59dd08820e0c2a9a7c3c9bf2231a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T08:02:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T08:02:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:02:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.527793 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0871bd641c5ad956258cafb9f56540f6fe98eacc7bbb22e4141d0464dc1a50c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.541091 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80e1945c1fafde38637ac19211e59e4296af067178859786ff6cfe274d5deff0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d657875f2de1149681421d14e0a70da5445711b8d00e9d876a6bf8293a7be8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.556665 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n85f4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5367714-1633-4695-af4e-4eb6419daa96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T08:04:05Z\\\",\\\"message\\\":\\\"2025-09-30T08:03:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292\\\\n2025-09-30T08:03:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_55b6e386-9aec-4ef3-86b7-1012f639b292 to /host/opt/cni/bin/\\\\n2025-09-30T08:03:20Z [verbose] multus-daemon started\\\\n2025-09-30T08:03:20Z [verbose] Readiness Indicator file check\\\\n2025-09-30T08:04:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsl8p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n85f4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.570253 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"136360fc-f06a-4b28-bbe6-b8cefcac4fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wdnx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pnxwm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.587085 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d6c52b207f924b1c96e2ee2a677dfa2b72f64ff20e427d587b479550f78b563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.592649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.592736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.592752 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.592772 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.592785 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:27Z","lastTransitionTime":"2025-09-30T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.606189 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.619667 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.631969 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86194959-30fe-4da8-a64b-f5a849c2f73a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T08:03:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://295878e719790b8c190386bd8f291084c7df06c65a68c6f3c310ff6ec8fb9fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T08:03:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffdsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T08:03:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hk2k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:27Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.695436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.695472 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.695484 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.695502 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.695514 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:27Z","lastTransitionTime":"2025-09-30T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.799112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.799184 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.799209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.799240 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.799312 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:27Z","lastTransitionTime":"2025-09-30T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.901597 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.901677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.901696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.901722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:27 crc kubenswrapper[4810]: I0930 08:04:27.901736 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:27Z","lastTransitionTime":"2025-09-30T08:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.004154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.004203 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.004219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.004239 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.004255 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:28Z","lastTransitionTime":"2025-09-30T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.106595 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.106639 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.106655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.106680 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.106697 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:28Z","lastTransitionTime":"2025-09-30T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.210351 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.210446 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.210470 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.210502 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.210525 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:28Z","lastTransitionTime":"2025-09-30T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.305889 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.305968 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:28 crc kubenswrapper[4810]: E0930 08:04:28.306231 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:28 crc kubenswrapper[4810]: E0930 08:04:28.306434 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.313312 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.313353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.313364 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.313382 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.313395 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:28Z","lastTransitionTime":"2025-09-30T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.415852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.415912 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.415931 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.415955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.415972 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:28Z","lastTransitionTime":"2025-09-30T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.518852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.518957 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.518976 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.518999 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.519020 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:28Z","lastTransitionTime":"2025-09-30T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.622181 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.622237 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.622258 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.622350 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.622376 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:28Z","lastTransitionTime":"2025-09-30T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.725131 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.725182 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.725199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.725223 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.725242 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:28Z","lastTransitionTime":"2025-09-30T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.827829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.827869 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.827885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.827906 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.827923 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:28Z","lastTransitionTime":"2025-09-30T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.931455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.931528 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.931547 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.931574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:28 crc kubenswrapper[4810]: I0930 08:04:28.931593 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:28Z","lastTransitionTime":"2025-09-30T08:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.035493 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.035549 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.035566 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.035594 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.035611 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:29Z","lastTransitionTime":"2025-09-30T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.138963 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.139022 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.139039 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.139065 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.139083 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:29Z","lastTransitionTime":"2025-09-30T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.242137 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.242185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.242200 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.242219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.242234 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:29Z","lastTransitionTime":"2025-09-30T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.306190 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.306237 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:29 crc kubenswrapper[4810]: E0930 08:04:29.306399 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:29 crc kubenswrapper[4810]: E0930 08:04:29.306779 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.345415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.345451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.345462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.345480 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.345492 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:29Z","lastTransitionTime":"2025-09-30T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.448151 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.448223 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.448243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.448274 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.448321 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:29Z","lastTransitionTime":"2025-09-30T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.551573 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.551644 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.551668 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.551698 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.551744 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:29Z","lastTransitionTime":"2025-09-30T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.655454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.655519 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.655538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.655572 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.655599 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:29Z","lastTransitionTime":"2025-09-30T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.759625 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.759719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.759745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.759778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.759804 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:29Z","lastTransitionTime":"2025-09-30T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.863340 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.863408 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.863429 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.863457 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.863476 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:29Z","lastTransitionTime":"2025-09-30T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.966768 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.966824 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.966845 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.966869 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:29 crc kubenswrapper[4810]: I0930 08:04:29.966886 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:29Z","lastTransitionTime":"2025-09-30T08:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.068614 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.068682 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.068696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.068716 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.068731 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:30Z","lastTransitionTime":"2025-09-30T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.171872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.171952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.171963 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.171981 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.171995 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:30Z","lastTransitionTime":"2025-09-30T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.275239 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.275313 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.275330 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.275355 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.275369 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:30Z","lastTransitionTime":"2025-09-30T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.305551 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:30 crc kubenswrapper[4810]: E0930 08:04:30.305693 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.305728 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:30 crc kubenswrapper[4810]: E0930 08:04:30.306478 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.379231 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.379356 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.379382 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.379452 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.379473 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:30Z","lastTransitionTime":"2025-09-30T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.482448 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.482551 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.482570 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.482599 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.482618 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:30Z","lastTransitionTime":"2025-09-30T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.585808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.585877 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.585894 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.585923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.585941 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:30Z","lastTransitionTime":"2025-09-30T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.690050 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.690114 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.690135 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.690163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.690182 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:30Z","lastTransitionTime":"2025-09-30T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.793211 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.793339 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.793358 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.793383 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.793406 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:30Z","lastTransitionTime":"2025-09-30T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.897054 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.897120 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.897138 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.897165 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:30 crc kubenswrapper[4810]: I0930 08:04:30.897182 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:30Z","lastTransitionTime":"2025-09-30T08:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.007349 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.007454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.007475 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.007501 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.007531 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:31Z","lastTransitionTime":"2025-09-30T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.111064 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.111128 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.111145 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.111170 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.111192 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:31Z","lastTransitionTime":"2025-09-30T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.213644 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.213688 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.213700 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.213717 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.213729 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:31Z","lastTransitionTime":"2025-09-30T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.305848 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.305903 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:31 crc kubenswrapper[4810]: E0930 08:04:31.306175 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:31 crc kubenswrapper[4810]: E0930 08:04:31.306322 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.316437 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.316487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.316505 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.316527 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.316547 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:31Z","lastTransitionTime":"2025-09-30T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.419699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.419756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.419775 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.419807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.419857 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:31Z","lastTransitionTime":"2025-09-30T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.524839 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.524894 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.524912 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.524934 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.524952 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:31Z","lastTransitionTime":"2025-09-30T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.628623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.628686 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.628704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.628730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.628752 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:31Z","lastTransitionTime":"2025-09-30T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.732633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.732702 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.732722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.732753 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.732773 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:31Z","lastTransitionTime":"2025-09-30T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.836366 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.836428 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.836440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.836461 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.836482 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:31Z","lastTransitionTime":"2025-09-30T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.939318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.939478 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.939508 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.939547 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:31 crc kubenswrapper[4810]: I0930 08:04:31.939571 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:31Z","lastTransitionTime":"2025-09-30T08:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.044301 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.044360 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.044377 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.044401 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.044421 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:32Z","lastTransitionTime":"2025-09-30T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.147221 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.147271 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.147307 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.147327 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.147343 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:32Z","lastTransitionTime":"2025-09-30T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.251152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.251228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.251246 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.251317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.251346 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:32Z","lastTransitionTime":"2025-09-30T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.305534 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.305642 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:32 crc kubenswrapper[4810]: E0930 08:04:32.306049 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:32 crc kubenswrapper[4810]: E0930 08:04:32.306151 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.355402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.355470 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.355495 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.355527 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.355548 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:32Z","lastTransitionTime":"2025-09-30T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.458700 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.458789 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.458818 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.458856 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.458883 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:32Z","lastTransitionTime":"2025-09-30T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.562223 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.562325 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.562345 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.562374 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.562393 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:32Z","lastTransitionTime":"2025-09-30T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.665025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.665080 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.665089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.665106 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.665117 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:32Z","lastTransitionTime":"2025-09-30T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.768390 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.768456 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.768473 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.768502 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.768519 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:32Z","lastTransitionTime":"2025-09-30T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.871915 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.871988 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.872023 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.872051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.872077 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:32Z","lastTransitionTime":"2025-09-30T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.975000 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.975077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.975100 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.975131 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:32 crc kubenswrapper[4810]: I0930 08:04:32.975155 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:32Z","lastTransitionTime":"2025-09-30T08:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.077582 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.077623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.077636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.077653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.077666 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:33Z","lastTransitionTime":"2025-09-30T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.180737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.180855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.180928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.180970 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.180994 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:33Z","lastTransitionTime":"2025-09-30T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.284933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.285035 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.285089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.285116 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.285134 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:33Z","lastTransitionTime":"2025-09-30T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.306429 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.306458 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:33 crc kubenswrapper[4810]: E0930 08:04:33.306651 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:33 crc kubenswrapper[4810]: E0930 08:04:33.306837 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.389328 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.389984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.390032 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.390064 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.390104 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:33Z","lastTransitionTime":"2025-09-30T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.494181 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.494295 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.494315 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.494341 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.494358 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:33Z","lastTransitionTime":"2025-09-30T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.598785 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.598859 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.598880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.598905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.598924 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:33Z","lastTransitionTime":"2025-09-30T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.701507 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.701576 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.701598 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.701624 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.701645 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:33Z","lastTransitionTime":"2025-09-30T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.804115 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.804144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.804152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.804167 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.804175 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:33Z","lastTransitionTime":"2025-09-30T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.906510 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.906571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.906588 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.906616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:33 crc kubenswrapper[4810]: I0930 08:04:33.906638 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:33Z","lastTransitionTime":"2025-09-30T08:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.010603 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.010684 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.010709 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.010740 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.010765 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:34Z","lastTransitionTime":"2025-09-30T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.113627 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.113720 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.113740 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.113774 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.113793 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:34Z","lastTransitionTime":"2025-09-30T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.217169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.217246 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.217297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.217327 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.217349 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:34Z","lastTransitionTime":"2025-09-30T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.306653 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.306758 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:34 crc kubenswrapper[4810]: E0930 08:04:34.306929 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:34 crc kubenswrapper[4810]: E0930 08:04:34.307096 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.321836 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.321900 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.321918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.321952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.321978 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:34Z","lastTransitionTime":"2025-09-30T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.425228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.425313 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.425332 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.425358 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.425376 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:34Z","lastTransitionTime":"2025-09-30T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.528797 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.528861 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.528879 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.528909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.528926 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:34Z","lastTransitionTime":"2025-09-30T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.632329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.632369 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.632380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.632399 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.632413 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:34Z","lastTransitionTime":"2025-09-30T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.735367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.735755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.735965 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.736132 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.736301 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:34Z","lastTransitionTime":"2025-09-30T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.839904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.839978 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.839999 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.840025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.840045 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:34Z","lastTransitionTime":"2025-09-30T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.943645 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.943693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.943705 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.943722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:34 crc kubenswrapper[4810]: I0930 08:04:34.943735 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:34Z","lastTransitionTime":"2025-09-30T08:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.046743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.046845 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.046872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.046905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.046929 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.149951 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.150053 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.150074 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.150102 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.150119 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.253323 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.253385 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.253406 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.253433 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.253450 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.306102 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.306188 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:35 crc kubenswrapper[4810]: E0930 08:04:35.306303 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:35 crc kubenswrapper[4810]: E0930 08:04:35.306654 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.354359 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.354415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.354432 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.354458 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.354478 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: E0930 08:04:35.377134 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:35Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.383081 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.383166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.383192 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.383229 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.383254 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: E0930 08:04:35.405867 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:35Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.411800 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.411862 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.411896 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.411956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.411981 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: E0930 08:04:35.434469 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:35Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.439865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.439902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.439918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.439938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.439952 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: E0930 08:04:35.453749 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:35Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.458736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.458777 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.458796 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.458819 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.458835 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: E0930 08:04:35.479454 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T08:04:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"441a0caf-7730-4ad7-a34e-ac02f226b5b0\\\",\\\"systemUUID\\\":\\\"3de9b6fb-1269-4c3c-a6c1-7b041590cd04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T08:04:35Z is after 2025-08-24T17:21:41Z" Sep 30 08:04:35 crc kubenswrapper[4810]: E0930 08:04:35.479830 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.481965 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.482030 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.482054 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.482084 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.482111 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.585863 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.585955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.585973 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.586007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.586026 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.689783 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.689855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.689878 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.689910 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.689934 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.793079 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.793154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.793190 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.793225 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.793247 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.896570 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.896650 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.896676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.896711 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:35 crc kubenswrapper[4810]: I0930 08:04:35.896736 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:35Z","lastTransitionTime":"2025-09-30T08:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.000506 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.000571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.000589 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.000617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.000635 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:36Z","lastTransitionTime":"2025-09-30T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.104119 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.104189 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.104207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.104235 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.104252 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:36Z","lastTransitionTime":"2025-09-30T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.208168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.208234 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.208259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.208367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.208393 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:36Z","lastTransitionTime":"2025-09-30T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.306019 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.306041 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:36 crc kubenswrapper[4810]: E0930 08:04:36.306657 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:36 crc kubenswrapper[4810]: E0930 08:04:36.306838 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.311145 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.311371 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.311571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.311782 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.312001 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:36Z","lastTransitionTime":"2025-09-30T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.320999 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:36 crc kubenswrapper[4810]: E0930 08:04:36.321216 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:04:36 crc kubenswrapper[4810]: E0930 08:04:36.321569 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs podName:136360fc-f06a-4b28-bbe6-b8cefcac4fda nodeName:}" failed. No retries permitted until 2025-09-30 08:05:40.321528707 +0000 UTC m=+163.773728014 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs") pod "network-metrics-daemon-pnxwm" (UID: "136360fc-f06a-4b28-bbe6-b8cefcac4fda") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.415232 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.415326 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.415337 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.415356 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.415367 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:36Z","lastTransitionTime":"2025-09-30T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.518902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.518987 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.519012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.519047 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.519072 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:36Z","lastTransitionTime":"2025-09-30T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.623018 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.623390 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.623577 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.623736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.623867 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:36Z","lastTransitionTime":"2025-09-30T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.726464 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.726524 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.726543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.726566 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.726583 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:36Z","lastTransitionTime":"2025-09-30T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.830163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.830221 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.830242 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.830296 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.830314 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:36Z","lastTransitionTime":"2025-09-30T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.933365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.933456 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.933482 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.933514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:36 crc kubenswrapper[4810]: I0930 08:04:36.933538 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:36Z","lastTransitionTime":"2025-09-30T08:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.036907 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.036969 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.036987 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.037017 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.037037 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:37Z","lastTransitionTime":"2025-09-30T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.140506 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.140561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.140585 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.140686 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.140711 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:37Z","lastTransitionTime":"2025-09-30T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.244515 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.244574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.244593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.244622 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.244644 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:37Z","lastTransitionTime":"2025-09-30T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.305569 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.305726 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:37 crc kubenswrapper[4810]: E0930 08:04:37.305892 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:37 crc kubenswrapper[4810]: E0930 08:04:37.306169 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.343778 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=47.343677751 podStartE2EDuration="47.343677751s" podCreationTimestamp="2025-09-30 08:03:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:37.343618709 +0000 UTC m=+100.795818016" watchObservedRunningTime="2025-09-30 08:04:37.343677751 +0000 UTC m=+100.795877058" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.351716 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.351811 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.351836 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.352507 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.352574 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:37Z","lastTransitionTime":"2025-09-30T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.442762 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-n85f4" podStartSLOduration=80.442724017 podStartE2EDuration="1m20.442724017s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:37.412495791 +0000 UTC m=+100.864695048" watchObservedRunningTime="2025-09-30 08:04:37.442724017 +0000 UTC m=+100.894923324" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.458079 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.458199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.458228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.458263 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.458321 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:37Z","lastTransitionTime":"2025-09-30T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.484618 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-fnzgz" podStartSLOduration=80.484522202 podStartE2EDuration="1m20.484522202s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:37.443465568 +0000 UTC m=+100.895664925" watchObservedRunningTime="2025-09-30 08:04:37.484522202 +0000 UTC m=+100.936721509" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.515658 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=80.515626573 podStartE2EDuration="1m20.515626573s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:37.513978206 +0000 UTC m=+100.966177513" watchObservedRunningTime="2025-09-30 08:04:37.515626573 +0000 UTC m=+100.967825870" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.547050 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=75.547007622 podStartE2EDuration="1m15.547007622s" podCreationTimestamp="2025-09-30 08:03:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:37.545772407 +0000 UTC m=+100.997971714" watchObservedRunningTime="2025-09-30 08:04:37.547007622 +0000 UTC m=+100.999206919" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.562183 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.562259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.562307 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.562338 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.562362 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:37Z","lastTransitionTime":"2025-09-30T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.638477 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podStartSLOduration=80.638449233 podStartE2EDuration="1m20.638449233s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:37.63832749 +0000 UTC m=+101.090526757" watchObservedRunningTime="2025-09-30 08:04:37.638449233 +0000 UTC m=+101.090648500" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.664471 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.664518 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.664530 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.664553 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.664568 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:37Z","lastTransitionTime":"2025-09-30T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.689918 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-xrbr6" podStartSLOduration=80.689898271 podStartE2EDuration="1m20.689898271s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:37.689872601 +0000 UTC m=+101.142071878" watchObservedRunningTime="2025-09-30 08:04:37.689898271 +0000 UTC m=+101.142097528" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.719609 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=10.719587593 podStartE2EDuration="10.719587593s" podCreationTimestamp="2025-09-30 08:04:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:37.706032618 +0000 UTC m=+101.158231895" watchObservedRunningTime="2025-09-30 08:04:37.719587593 +0000 UTC m=+101.171786860" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.730332 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jlk85" podStartSLOduration=79.730309076 podStartE2EDuration="1m19.730309076s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:37.729878054 +0000 UTC m=+101.182077331" watchObservedRunningTime="2025-09-30 08:04:37.730309076 +0000 UTC m=+101.182508343" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.756201 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=77.75618781 podStartE2EDuration="1m17.75618781s" podCreationTimestamp="2025-09-30 08:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:37.755992564 +0000 UTC m=+101.208191841" watchObservedRunningTime="2025-09-30 08:04:37.75618781 +0000 UTC m=+101.208387077" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.765915 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-z5md6" podStartSLOduration=80.765891425 podStartE2EDuration="1m20.765891425s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:37.764928607 +0000 UTC m=+101.217127884" watchObservedRunningTime="2025-09-30 08:04:37.765891425 +0000 UTC m=+101.218090702" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.766524 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.766564 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.766574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.766592 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.766602 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:37Z","lastTransitionTime":"2025-09-30T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.869077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.869132 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.869147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.869163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.869172 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:37Z","lastTransitionTime":"2025-09-30T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.971978 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.972055 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.972077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.972109 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:37 crc kubenswrapper[4810]: I0930 08:04:37.972130 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:37Z","lastTransitionTime":"2025-09-30T08:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.075396 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.075451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.075464 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.075486 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.075501 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:38Z","lastTransitionTime":"2025-09-30T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.178638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.178712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.178726 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.178754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.178771 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:38Z","lastTransitionTime":"2025-09-30T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.282078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.282145 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.282169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.282201 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.282222 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:38Z","lastTransitionTime":"2025-09-30T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.305956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.305975 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:38 crc kubenswrapper[4810]: E0930 08:04:38.306157 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:38 crc kubenswrapper[4810]: E0930 08:04:38.306258 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.385397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.385484 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.385503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.385530 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.385550 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:38Z","lastTransitionTime":"2025-09-30T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.489172 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.489239 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.489255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.489309 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.489330 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:38Z","lastTransitionTime":"2025-09-30T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.592537 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.592602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.592619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.592647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.592666 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:38Z","lastTransitionTime":"2025-09-30T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.696581 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.696644 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.696661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.696685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.696703 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:38Z","lastTransitionTime":"2025-09-30T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.800582 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.800664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.800684 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.800710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.800730 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:38Z","lastTransitionTime":"2025-09-30T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.903977 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.904071 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.904096 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.904130 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:38 crc kubenswrapper[4810]: I0930 08:04:38.904156 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:38Z","lastTransitionTime":"2025-09-30T08:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.007168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.007525 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.007541 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.007565 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.007598 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:39Z","lastTransitionTime":"2025-09-30T08:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.111169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.111221 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.111241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.111300 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.111320 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:39Z","lastTransitionTime":"2025-09-30T08:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.214756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.214812 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.214829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.214855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.214874 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:39Z","lastTransitionTime":"2025-09-30T08:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.306571 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.306676 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:39 crc kubenswrapper[4810]: E0930 08:04:39.306799 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:39 crc kubenswrapper[4810]: E0930 08:04:39.306886 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.317321 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.317364 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.317380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.317401 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.317417 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:39Z","lastTransitionTime":"2025-09-30T08:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.420600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.420652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.420665 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.420682 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.420694 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:39Z","lastTransitionTime":"2025-09-30T08:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.523676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.523738 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.523756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.523788 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.523806 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:39Z","lastTransitionTime":"2025-09-30T08:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.626178 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.626237 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.626254 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.626308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.626327 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:39Z","lastTransitionTime":"2025-09-30T08:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.729169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.729227 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.729244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.729299 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.729316 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:39Z","lastTransitionTime":"2025-09-30T08:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.831434 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.831482 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.831496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.831518 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.831532 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:39Z","lastTransitionTime":"2025-09-30T08:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.934748 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.934815 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.934838 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.934870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:39 crc kubenswrapper[4810]: I0930 08:04:39.934894 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:39Z","lastTransitionTime":"2025-09-30T08:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.037773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.037845 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.037870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.037901 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.037924 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:40Z","lastTransitionTime":"2025-09-30T08:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.141338 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.141423 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.141443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.141473 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.141496 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:40Z","lastTransitionTime":"2025-09-30T08:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.244725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.244782 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.244799 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.244824 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.244840 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:40Z","lastTransitionTime":"2025-09-30T08:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.306331 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.306393 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:40 crc kubenswrapper[4810]: E0930 08:04:40.306514 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:40 crc kubenswrapper[4810]: E0930 08:04:40.306675 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.347817 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.347880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.347898 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.347922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.347940 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:40Z","lastTransitionTime":"2025-09-30T08:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.451628 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.451676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.451694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.451717 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.451734 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:40Z","lastTransitionTime":"2025-09-30T08:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.554564 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.554624 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.554641 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.554664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.554681 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:40Z","lastTransitionTime":"2025-09-30T08:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.659053 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.659113 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.659131 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.659157 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.659177 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:40Z","lastTransitionTime":"2025-09-30T08:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.762757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.762813 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.762831 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.762855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.762874 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:40Z","lastTransitionTime":"2025-09-30T08:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.865758 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.865825 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.865842 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.865867 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.865885 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:40Z","lastTransitionTime":"2025-09-30T08:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.968680 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.968743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.968756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.968777 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:40 crc kubenswrapper[4810]: I0930 08:04:40.968790 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:40Z","lastTransitionTime":"2025-09-30T08:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.071861 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.071935 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.071959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.071989 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.072006 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:41Z","lastTransitionTime":"2025-09-30T08:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.174892 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.174985 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.175007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.175034 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.175056 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:41Z","lastTransitionTime":"2025-09-30T08:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.278672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.278728 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.278747 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.278772 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.278789 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:41Z","lastTransitionTime":"2025-09-30T08:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.306301 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.306502 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:41 crc kubenswrapper[4810]: E0930 08:04:41.306652 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:41 crc kubenswrapper[4810]: E0930 08:04:41.306860 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.382045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.382086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.382097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.382114 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.382125 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:41Z","lastTransitionTime":"2025-09-30T08:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.485388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.485453 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.485473 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.485497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.485515 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:41Z","lastTransitionTime":"2025-09-30T08:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.588538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.588613 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.588635 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.588664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.588717 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:41Z","lastTransitionTime":"2025-09-30T08:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.692372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.692446 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.692462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.692482 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.692499 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:41Z","lastTransitionTime":"2025-09-30T08:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.796028 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.796099 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.796116 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.796144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.796162 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:41Z","lastTransitionTime":"2025-09-30T08:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.899761 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.899824 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.899847 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.899874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:41 crc kubenswrapper[4810]: I0930 08:04:41.899892 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:41Z","lastTransitionTime":"2025-09-30T08:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.003261 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.003340 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.003353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.003372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.003389 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:42Z","lastTransitionTime":"2025-09-30T08:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.106220 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.106323 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.106352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.106380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.106399 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:42Z","lastTransitionTime":"2025-09-30T08:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.209167 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.209241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.209263 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.209342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.209365 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:42Z","lastTransitionTime":"2025-09-30T08:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.306524 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.307443 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:42 crc kubenswrapper[4810]: E0930 08:04:42.308086 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:42 crc kubenswrapper[4810]: E0930 08:04:42.308197 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.308546 4810 scope.go:117] "RemoveContainer" containerID="08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b" Sep 30 08:04:42 crc kubenswrapper[4810]: E0930 08:04:42.308813 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9bkcj_openshift-ovn-kubernetes(aacbcdc8-7ef5-473b-8055-145f40040d5d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.312509 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.312563 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.312596 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.312638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.312664 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:42Z","lastTransitionTime":"2025-09-30T08:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.415778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.415840 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.415857 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.415883 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.415902 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:42Z","lastTransitionTime":"2025-09-30T08:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.519840 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.519922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.519947 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.519981 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.520006 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:42Z","lastTransitionTime":"2025-09-30T08:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.624170 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.624244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.624303 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.624339 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.624359 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:42Z","lastTransitionTime":"2025-09-30T08:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.727993 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.728061 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.728083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.728111 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.728129 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:42Z","lastTransitionTime":"2025-09-30T08:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.831712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.831782 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.831807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.831839 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.831863 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:42Z","lastTransitionTime":"2025-09-30T08:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.935234 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.935354 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.935379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.935416 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:42 crc kubenswrapper[4810]: I0930 08:04:42.935436 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:42Z","lastTransitionTime":"2025-09-30T08:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.038492 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.038568 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.038592 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.038623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.038645 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:43Z","lastTransitionTime":"2025-09-30T08:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.142443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.142534 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.142552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.142578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.142596 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:43Z","lastTransitionTime":"2025-09-30T08:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.245943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.246009 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.246027 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.246054 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.246073 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:43Z","lastTransitionTime":"2025-09-30T08:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.305880 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.305903 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:43 crc kubenswrapper[4810]: E0930 08:04:43.306121 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:43 crc kubenswrapper[4810]: E0930 08:04:43.306359 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.348364 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.348437 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.348460 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.348489 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.348510 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:43Z","lastTransitionTime":"2025-09-30T08:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.454680 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.454744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.454761 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.454788 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.454800 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:43Z","lastTransitionTime":"2025-09-30T08:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.558366 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.558444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.558468 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.558502 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.558526 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:43Z","lastTransitionTime":"2025-09-30T08:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.662938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.663012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.663031 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.663633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.663707 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:43Z","lastTransitionTime":"2025-09-30T08:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.767545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.767612 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.767635 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.767666 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.767688 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:43Z","lastTransitionTime":"2025-09-30T08:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.870687 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.870772 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.870798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.870830 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.870852 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:43Z","lastTransitionTime":"2025-09-30T08:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.974369 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.974444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.974471 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.974501 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:43 crc kubenswrapper[4810]: I0930 08:04:43.974522 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:43Z","lastTransitionTime":"2025-09-30T08:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.077465 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.077545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.077573 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.077604 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.077628 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:44Z","lastTransitionTime":"2025-09-30T08:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.181700 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.181767 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.181786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.181812 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.181829 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:44Z","lastTransitionTime":"2025-09-30T08:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.284662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.284724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.284741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.284766 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.284785 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:44Z","lastTransitionTime":"2025-09-30T08:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.305914 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.305915 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:44 crc kubenswrapper[4810]: E0930 08:04:44.306096 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:44 crc kubenswrapper[4810]: E0930 08:04:44.306186 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.387713 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.387764 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.387782 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.387806 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.387824 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:44Z","lastTransitionTime":"2025-09-30T08:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.490818 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.490868 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.490886 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.490908 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.490929 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:44Z","lastTransitionTime":"2025-09-30T08:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.594945 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.595019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.595040 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.595069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.595095 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:44Z","lastTransitionTime":"2025-09-30T08:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.697382 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.697443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.697460 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.697483 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.697500 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:44Z","lastTransitionTime":"2025-09-30T08:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.800870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.800937 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.800959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.800989 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.801016 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:44Z","lastTransitionTime":"2025-09-30T08:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.903878 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.904191 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.904216 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.904243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:44 crc kubenswrapper[4810]: I0930 08:04:44.904262 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:44Z","lastTransitionTime":"2025-09-30T08:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.007476 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.007568 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.007586 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.007612 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.007631 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:45Z","lastTransitionTime":"2025-09-30T08:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.110744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.110804 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.110820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.110844 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.110862 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:45Z","lastTransitionTime":"2025-09-30T08:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.213831 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.213902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.213925 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.213959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.213983 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:45Z","lastTransitionTime":"2025-09-30T08:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.306165 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:45 crc kubenswrapper[4810]: E0930 08:04:45.306375 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.306183 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:45 crc kubenswrapper[4810]: E0930 08:04:45.306625 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.316616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.316668 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.316685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.316711 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.316727 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:45Z","lastTransitionTime":"2025-09-30T08:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.419901 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.419980 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.419998 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.420027 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.420049 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:45Z","lastTransitionTime":"2025-09-30T08:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.523735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.523795 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.523806 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.523844 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.523858 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:45Z","lastTransitionTime":"2025-09-30T08:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.627120 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.627195 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.627212 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.627237 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.627260 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:45Z","lastTransitionTime":"2025-09-30T08:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.730412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.730485 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.730503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.730539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.730560 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:45Z","lastTransitionTime":"2025-09-30T08:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.742841 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.742904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.742927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.742956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.742973 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T08:04:45Z","lastTransitionTime":"2025-09-30T08:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.816223 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t"] Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.816767 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.818578 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.819465 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.820098 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.820239 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.933996 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.934102 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.934244 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.934335 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:45 crc kubenswrapper[4810]: I0930 08:04:45.934498 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.035766 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.035863 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.035951 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.036077 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.036118 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.036207 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.036302 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.037552 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.054476 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.070201 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8afac2ef-bda5-483d-9fc2-ac18928bb1c2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wbv2t\" (UID: \"8afac2ef-bda5-483d-9fc2-ac18928bb1c2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.144153 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.305779 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:46 crc kubenswrapper[4810]: I0930 08:04:46.305810 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:46 crc kubenswrapper[4810]: E0930 08:04:46.305978 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:46 crc kubenswrapper[4810]: E0930 08:04:46.306118 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:47 crc kubenswrapper[4810]: I0930 08:04:47.047337 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" event={"ID":"8afac2ef-bda5-483d-9fc2-ac18928bb1c2","Type":"ContainerStarted","Data":"811c2be6b963c853a79940de46b20bb7fbf14949ca5b236398c7608f5c234539"} Sep 30 08:04:47 crc kubenswrapper[4810]: I0930 08:04:47.047440 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" event={"ID":"8afac2ef-bda5-483d-9fc2-ac18928bb1c2","Type":"ContainerStarted","Data":"efb524c420a3a9b54918824d5bdabd72352e7f61ba7bd1fc23fc48a717d7f959"} Sep 30 08:04:47 crc kubenswrapper[4810]: I0930 08:04:47.306378 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:47 crc kubenswrapper[4810]: I0930 08:04:47.306402 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:47 crc kubenswrapper[4810]: E0930 08:04:47.308544 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:47 crc kubenswrapper[4810]: E0930 08:04:47.308636 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:48 crc kubenswrapper[4810]: I0930 08:04:48.305867 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:48 crc kubenswrapper[4810]: E0930 08:04:48.306075 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:48 crc kubenswrapper[4810]: I0930 08:04:48.306851 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:48 crc kubenswrapper[4810]: E0930 08:04:48.307168 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:49 crc kubenswrapper[4810]: I0930 08:04:49.305985 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:49 crc kubenswrapper[4810]: I0930 08:04:49.306041 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:49 crc kubenswrapper[4810]: E0930 08:04:49.306233 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:49 crc kubenswrapper[4810]: E0930 08:04:49.306447 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:50 crc kubenswrapper[4810]: I0930 08:04:50.305924 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:50 crc kubenswrapper[4810]: I0930 08:04:50.305934 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:50 crc kubenswrapper[4810]: E0930 08:04:50.306125 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:50 crc kubenswrapper[4810]: E0930 08:04:50.306306 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:51 crc kubenswrapper[4810]: I0930 08:04:51.306604 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:51 crc kubenswrapper[4810]: I0930 08:04:51.306763 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:51 crc kubenswrapper[4810]: E0930 08:04:51.306985 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:51 crc kubenswrapper[4810]: E0930 08:04:51.307108 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:52 crc kubenswrapper[4810]: I0930 08:04:52.068198 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n85f4_f5367714-1633-4695-af4e-4eb6419daa96/kube-multus/1.log" Sep 30 08:04:52 crc kubenswrapper[4810]: I0930 08:04:52.069021 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n85f4_f5367714-1633-4695-af4e-4eb6419daa96/kube-multus/0.log" Sep 30 08:04:52 crc kubenswrapper[4810]: I0930 08:04:52.069095 4810 generic.go:334] "Generic (PLEG): container finished" podID="f5367714-1633-4695-af4e-4eb6419daa96" containerID="535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857" exitCode=1 Sep 30 08:04:52 crc kubenswrapper[4810]: I0930 08:04:52.069138 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n85f4" event={"ID":"f5367714-1633-4695-af4e-4eb6419daa96","Type":"ContainerDied","Data":"535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857"} Sep 30 08:04:52 crc kubenswrapper[4810]: I0930 08:04:52.069188 4810 scope.go:117] "RemoveContainer" containerID="de0a03ed304d970f9867c4e20a64827c4899eba2846ed9db4335b0b3f4edaeab" Sep 30 08:04:52 crc kubenswrapper[4810]: I0930 08:04:52.069760 4810 scope.go:117] "RemoveContainer" containerID="535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857" Sep 30 08:04:52 crc kubenswrapper[4810]: E0930 08:04:52.070047 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-n85f4_openshift-multus(f5367714-1633-4695-af4e-4eb6419daa96)\"" pod="openshift-multus/multus-n85f4" podUID="f5367714-1633-4695-af4e-4eb6419daa96" Sep 30 08:04:52 crc kubenswrapper[4810]: I0930 08:04:52.102075 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wbv2t" podStartSLOduration=95.102049804 podStartE2EDuration="1m35.102049804s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:47.07076024 +0000 UTC m=+110.522959547" watchObservedRunningTime="2025-09-30 08:04:52.102049804 +0000 UTC m=+115.554249111" Sep 30 08:04:52 crc kubenswrapper[4810]: I0930 08:04:52.305366 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:52 crc kubenswrapper[4810]: I0930 08:04:52.305536 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:52 crc kubenswrapper[4810]: E0930 08:04:52.305700 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:52 crc kubenswrapper[4810]: E0930 08:04:52.305925 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:53 crc kubenswrapper[4810]: I0930 08:04:53.075476 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n85f4_f5367714-1633-4695-af4e-4eb6419daa96/kube-multus/1.log" Sep 30 08:04:53 crc kubenswrapper[4810]: I0930 08:04:53.305621 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:53 crc kubenswrapper[4810]: E0930 08:04:53.305826 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:53 crc kubenswrapper[4810]: I0930 08:04:53.305911 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:53 crc kubenswrapper[4810]: E0930 08:04:53.306082 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:54 crc kubenswrapper[4810]: I0930 08:04:54.305502 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:54 crc kubenswrapper[4810]: I0930 08:04:54.305536 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:54 crc kubenswrapper[4810]: E0930 08:04:54.305767 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:54 crc kubenswrapper[4810]: E0930 08:04:54.305859 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:55 crc kubenswrapper[4810]: I0930 08:04:55.306257 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:55 crc kubenswrapper[4810]: I0930 08:04:55.306450 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:55 crc kubenswrapper[4810]: E0930 08:04:55.307498 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:55 crc kubenswrapper[4810]: E0930 08:04:55.307319 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:56 crc kubenswrapper[4810]: I0930 08:04:56.305631 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:56 crc kubenswrapper[4810]: I0930 08:04:56.305695 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:56 crc kubenswrapper[4810]: E0930 08:04:56.305866 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:56 crc kubenswrapper[4810]: E0930 08:04:56.305978 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:57 crc kubenswrapper[4810]: E0930 08:04:57.235783 4810 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Sep 30 08:04:57 crc kubenswrapper[4810]: I0930 08:04:57.306387 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:57 crc kubenswrapper[4810]: I0930 08:04:57.306529 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:57 crc kubenswrapper[4810]: E0930 08:04:57.309048 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:57 crc kubenswrapper[4810]: E0930 08:04:57.309368 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:04:57 crc kubenswrapper[4810]: I0930 08:04:57.310497 4810 scope.go:117] "RemoveContainer" containerID="08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b" Sep 30 08:04:57 crc kubenswrapper[4810]: E0930 08:04:57.408861 4810 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 08:04:58 crc kubenswrapper[4810]: I0930 08:04:58.098685 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/3.log" Sep 30 08:04:58 crc kubenswrapper[4810]: I0930 08:04:58.102708 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerStarted","Data":"933fa6a06cdfcff44c3ae380d64b817e01623656491d875120a01cdfbdbdff55"} Sep 30 08:04:58 crc kubenswrapper[4810]: I0930 08:04:58.103303 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:04:58 crc kubenswrapper[4810]: I0930 08:04:58.305841 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:04:58 crc kubenswrapper[4810]: I0930 08:04:58.305956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:58 crc kubenswrapper[4810]: E0930 08:04:58.305971 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:04:58 crc kubenswrapper[4810]: E0930 08:04:58.306246 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:58 crc kubenswrapper[4810]: I0930 08:04:58.309490 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podStartSLOduration=101.309478221 podStartE2EDuration="1m41.309478221s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:04:58.147070429 +0000 UTC m=+121.599269706" watchObservedRunningTime="2025-09-30 08:04:58.309478221 +0000 UTC m=+121.761677488" Sep 30 08:04:58 crc kubenswrapper[4810]: I0930 08:04:58.310137 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-pnxwm"] Sep 30 08:04:59 crc kubenswrapper[4810]: I0930 08:04:59.106440 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:04:59 crc kubenswrapper[4810]: E0930 08:04:59.106663 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:04:59 crc kubenswrapper[4810]: I0930 08:04:59.306540 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:04:59 crc kubenswrapper[4810]: I0930 08:04:59.306725 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:04:59 crc kubenswrapper[4810]: E0930 08:04:59.306740 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:04:59 crc kubenswrapper[4810]: E0930 08:04:59.306934 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:05:00 crc kubenswrapper[4810]: I0930 08:05:00.306233 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:05:00 crc kubenswrapper[4810]: E0930 08:05:00.306454 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:05:00 crc kubenswrapper[4810]: I0930 08:05:00.306233 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:05:00 crc kubenswrapper[4810]: E0930 08:05:00.306635 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:05:01 crc kubenswrapper[4810]: I0930 08:05:01.306296 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:01 crc kubenswrapper[4810]: I0930 08:05:01.306340 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:05:01 crc kubenswrapper[4810]: E0930 08:05:01.306446 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:05:01 crc kubenswrapper[4810]: E0930 08:05:01.306569 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:05:02 crc kubenswrapper[4810]: I0930 08:05:02.306494 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:05:02 crc kubenswrapper[4810]: I0930 08:05:02.306540 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:05:02 crc kubenswrapper[4810]: E0930 08:05:02.306713 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:05:02 crc kubenswrapper[4810]: E0930 08:05:02.306911 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:05:02 crc kubenswrapper[4810]: E0930 08:05:02.410297 4810 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 08:05:03 crc kubenswrapper[4810]: I0930 08:05:03.306180 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:05:03 crc kubenswrapper[4810]: E0930 08:05:03.306415 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:05:03 crc kubenswrapper[4810]: I0930 08:05:03.306489 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:03 crc kubenswrapper[4810]: E0930 08:05:03.306721 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:05:04 crc kubenswrapper[4810]: I0930 08:05:04.306116 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:05:04 crc kubenswrapper[4810]: I0930 08:05:04.306188 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:05:04 crc kubenswrapper[4810]: E0930 08:05:04.306360 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:05:04 crc kubenswrapper[4810]: E0930 08:05:04.306539 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:05:05 crc kubenswrapper[4810]: I0930 08:05:05.305761 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:05:05 crc kubenswrapper[4810]: I0930 08:05:05.305860 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:05 crc kubenswrapper[4810]: E0930 08:05:05.306069 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:05:05 crc kubenswrapper[4810]: E0930 08:05:05.306207 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:05:06 crc kubenswrapper[4810]: I0930 08:05:06.305976 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:05:06 crc kubenswrapper[4810]: I0930 08:05:06.305931 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:05:06 crc kubenswrapper[4810]: E0930 08:05:06.306409 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:05:06 crc kubenswrapper[4810]: E0930 08:05:06.306383 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:05:07 crc kubenswrapper[4810]: I0930 08:05:07.305417 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:07 crc kubenswrapper[4810]: I0930 08:05:07.305491 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:05:07 crc kubenswrapper[4810]: E0930 08:05:07.307620 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:05:07 crc kubenswrapper[4810]: E0930 08:05:07.307785 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:05:07 crc kubenswrapper[4810]: I0930 08:05:07.308484 4810 scope.go:117] "RemoveContainer" containerID="535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857" Sep 30 08:05:07 crc kubenswrapper[4810]: E0930 08:05:07.411348 4810 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 08:05:08 crc kubenswrapper[4810]: I0930 08:05:08.146783 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n85f4_f5367714-1633-4695-af4e-4eb6419daa96/kube-multus/1.log" Sep 30 08:05:08 crc kubenswrapper[4810]: I0930 08:05:08.146897 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n85f4" event={"ID":"f5367714-1633-4695-af4e-4eb6419daa96","Type":"ContainerStarted","Data":"8a0c05c5c70b22d510babd5c578400d035ba077b50083c564906b0e8dc68be45"} Sep 30 08:05:08 crc kubenswrapper[4810]: I0930 08:05:08.305412 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:05:08 crc kubenswrapper[4810]: I0930 08:05:08.305475 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:05:08 crc kubenswrapper[4810]: E0930 08:05:08.305590 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:05:08 crc kubenswrapper[4810]: E0930 08:05:08.306078 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:05:09 crc kubenswrapper[4810]: I0930 08:05:09.305918 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:05:09 crc kubenswrapper[4810]: I0930 08:05:09.305934 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:09 crc kubenswrapper[4810]: E0930 08:05:09.306107 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:05:09 crc kubenswrapper[4810]: E0930 08:05:09.306198 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:05:10 crc kubenswrapper[4810]: I0930 08:05:10.305990 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:05:10 crc kubenswrapper[4810]: I0930 08:05:10.306089 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:05:10 crc kubenswrapper[4810]: E0930 08:05:10.306210 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:05:10 crc kubenswrapper[4810]: E0930 08:05:10.306419 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:05:11 crc kubenswrapper[4810]: I0930 08:05:11.305845 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:05:11 crc kubenswrapper[4810]: I0930 08:05:11.305843 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:11 crc kubenswrapper[4810]: E0930 08:05:11.306056 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 08:05:11 crc kubenswrapper[4810]: E0930 08:05:11.306344 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 08:05:12 crc kubenswrapper[4810]: I0930 08:05:12.305989 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:05:12 crc kubenswrapper[4810]: I0930 08:05:12.306045 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:05:12 crc kubenswrapper[4810]: E0930 08:05:12.306246 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pnxwm" podUID="136360fc-f06a-4b28-bbe6-b8cefcac4fda" Sep 30 08:05:12 crc kubenswrapper[4810]: E0930 08:05:12.306439 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 08:05:13 crc kubenswrapper[4810]: I0930 08:05:13.305887 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:05:13 crc kubenswrapper[4810]: I0930 08:05:13.305933 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:13 crc kubenswrapper[4810]: I0930 08:05:13.308986 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Sep 30 08:05:13 crc kubenswrapper[4810]: I0930 08:05:13.309074 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Sep 30 08:05:13 crc kubenswrapper[4810]: I0930 08:05:13.309076 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Sep 30 08:05:13 crc kubenswrapper[4810]: I0930 08:05:13.309164 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Sep 30 08:05:14 crc kubenswrapper[4810]: I0930 08:05:14.306289 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:05:14 crc kubenswrapper[4810]: I0930 08:05:14.306304 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:05:14 crc kubenswrapper[4810]: I0930 08:05:14.311572 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Sep 30 08:05:14 crc kubenswrapper[4810]: I0930 08:05:14.311655 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.812600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.872135 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.875192 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.876535 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qdcml"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.877368 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.881052 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pq7h8"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.882488 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.883730 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nj7lz"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.884473 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.894914 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.918489 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.919595 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.920165 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nzjhz"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.920640 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.921356 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.923225 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.924368 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.924520 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.925804 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.925832 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.925947 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.925992 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926103 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926181 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926216 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926215 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926324 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926409 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926531 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926570 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926679 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926764 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926783 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926872 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926931 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.926983 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.927030 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.927083 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.927131 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.927255 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.927588 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.927708 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.927930 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.928224 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.928330 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-47kbw"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.928896 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-47kbw" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.933135 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.933790 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.934129 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.934138 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.934344 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.934409 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.934780 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.934900 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.934352 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.935636 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.935802 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.935962 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.937763 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.939702 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.940041 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.940222 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.940386 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.940569 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.940750 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.940861 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.941035 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.941130 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.941149 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.941191 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.941388 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.941454 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.942039 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.942421 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.942619 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.942750 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.942763 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.942909 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.942903 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.944064 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.944541 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.944798 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4bvx"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.945415 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.945963 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.948329 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.948556 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-49gdq"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.949044 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.949408 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.949532 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.949795 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.950087 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.950302 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-5w577"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.950861 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.951605 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-smwhg"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.952300 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.968215 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.968491 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.969102 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.971000 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.971531 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.971797 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.971830 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.972079 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.972133 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.972289 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.972621 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj"] Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.972929 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.973152 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.977548 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Sep 30 08:05:16 crc kubenswrapper[4810]: I0930 08:05:16.989527 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.008639 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.009332 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cd62560a-d47e-49f2-b3cd-06568dc86f43-encryption-config\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.009377 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p895k\" (UniqueName: \"kubernetes.io/projected/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-kube-api-access-p895k\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.009409 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-encryption-config\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.009426 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk7wv\" (UniqueName: \"kubernetes.io/projected/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-kube-api-access-mk7wv\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.009467 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-client-ca\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.009492 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.009542 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-audit\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.009561 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-config\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.009584 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqp5h\" (UniqueName: \"kubernetes.io/projected/cd62560a-d47e-49f2-b3cd-06568dc86f43-kube-api-access-rqp5h\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.011843 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-p8cxw"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.012382 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.012608 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zhxt6"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.012675 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.012933 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013185 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013409 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013506 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cd62560a-d47e-49f2-b3cd-06568dc86f43-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013551 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-node-pullsecrets\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013579 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzj7l\" (UniqueName: \"kubernetes.io/projected/80a2e9d7-a1ec-436d-8634-89c763644dec-kube-api-access-gzj7l\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013604 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-etcd-client\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013625 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-config\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013640 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cd62560a-d47e-49f2-b3cd-06568dc86f43-audit-policies\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013656 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cd62560a-d47e-49f2-b3cd-06568dc86f43-audit-dir\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013673 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-config\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013703 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd62560a-d47e-49f2-b3cd-06568dc86f43-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013719 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013742 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-etcd-serving-ca\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013758 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-image-import-ca\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013776 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-serving-cert\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013814 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd62560a-d47e-49f2-b3cd-06568dc86f43-serving-cert\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013833 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-images\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013849 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-audit-dir\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013870 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cd62560a-d47e-49f2-b3cd-06568dc86f43-etcd-client\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013897 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a2e9d7-a1ec-436d-8634-89c763644dec-serving-cert\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.013969 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014055 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014154 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014243 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014414 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014433 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014496 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014585 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014627 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014657 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014783 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014888 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.014961 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.015487 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.015789 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.015808 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.015959 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.016087 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.016089 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.017228 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5lb54"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.018075 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-555gv"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.018491 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.018725 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.023707 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.024562 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.024908 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.026178 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.028166 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.032504 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.034626 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.035335 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.035670 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.035715 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.035993 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.036461 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.037961 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.038317 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.039070 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.043333 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.044133 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.044132 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.044733 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.047244 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.048032 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.050838 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.051640 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qdcml"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.051850 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.052220 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.052662 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.094323 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.094760 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.100344 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.108409 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.109937 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.109933 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.114381 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.114756 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.115312 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.116110 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.116835 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-client-ca\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.116872 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-encryption-config\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.116911 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk7wv\" (UniqueName: \"kubernetes.io/projected/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-kube-api-access-mk7wv\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.116935 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.116965 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-serving-cert\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117185 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-audit\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117284 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-config\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117320 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-oauth-config\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117362 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqp5h\" (UniqueName: \"kubernetes.io/projected/cd62560a-d47e-49f2-b3cd-06568dc86f43-kube-api-access-rqp5h\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117392 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31e65ba1-328b-4d4b-905c-6074ab5d1427-serving-cert\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117431 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6j99\" (UniqueName: \"kubernetes.io/projected/3a1774b8-c928-4843-85ba-e02066fd0f5c-kube-api-access-c6j99\") pod \"cluster-samples-operator-665b6dd947-jcsj5\" (UID: \"3a1774b8-c928-4843-85ba-e02066fd0f5c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117483 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cd62560a-d47e-49f2-b3cd-06568dc86f43-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117610 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8szn\" (UniqueName: \"kubernetes.io/projected/52e8e292-0788-4957-ac33-083fadb0f61b-kube-api-access-b8szn\") pod \"dns-operator-744455d44c-p8cxw\" (UID: \"52e8e292-0788-4957-ac33-083fadb0f61b\") " pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117670 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-config\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117726 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-node-pullsecrets\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117804 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgf9m\" (UniqueName: \"kubernetes.io/projected/31e65ba1-328b-4d4b-905c-6074ab5d1427-kube-api-access-vgf9m\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117842 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-trusted-ca-bundle\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117872 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzj7l\" (UniqueName: \"kubernetes.io/projected/80a2e9d7-a1ec-436d-8634-89c763644dec-kube-api-access-gzj7l\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117898 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-etcd-client\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117960 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-config\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.117985 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c807324c-da31-4d36-8640-186d9cc31097-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118158 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-config\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118216 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cd62560a-d47e-49f2-b3cd-06568dc86f43-audit-policies\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118300 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cd62560a-d47e-49f2-b3cd-06568dc86f43-audit-dir\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118439 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c807324c-da31-4d36-8640-186d9cc31097-metrics-tls\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118471 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cbfn\" (UniqueName: \"kubernetes.io/projected/64fd38ac-d0ef-430c-aed5-5a4e4944986c-kube-api-access-2cbfn\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118626 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-service-ca\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118680 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd62560a-d47e-49f2-b3cd-06568dc86f43-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118706 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118765 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-serving-cert\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118804 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/52e8e292-0788-4957-ac33-083fadb0f61b-metrics-tls\") pod \"dns-operator-744455d44c-p8cxw\" (UID: \"52e8e292-0788-4957-ac33-083fadb0f61b\") " pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118873 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-etcd-serving-ca\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118899 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-image-import-ca\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118934 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd62560a-d47e-49f2-b3cd-06568dc86f43-serving-cert\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118959 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-images\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.118987 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/31e65ba1-328b-4d4b-905c-6074ab5d1427-trusted-ca\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119006 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/479ebf74-62ea-4516-8c29-8d9cdf128fb4-serving-cert\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119155 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx9zz\" (UniqueName: \"kubernetes.io/projected/c807324c-da31-4d36-8640-186d9cc31097-kube-api-access-vx9zz\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119202 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-audit-dir\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119245 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31e65ba1-328b-4d4b-905c-6074ab5d1427-config\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119298 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cd62560a-d47e-49f2-b3cd-06568dc86f43-etcd-client\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119328 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wskjk\" (UniqueName: \"kubernetes.io/projected/479ebf74-62ea-4516-8c29-8d9cdf128fb4-kube-api-access-wskjk\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119403 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a2e9d7-a1ec-436d-8634-89c763644dec-serving-cert\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119510 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-config\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119568 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c807324c-da31-4d36-8640-186d9cc31097-trusted-ca\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119649 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cd62560a-d47e-49f2-b3cd-06568dc86f43-encryption-config\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119673 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119696 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p895k\" (UniqueName: \"kubernetes.io/projected/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-kube-api-access-p895k\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119718 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-client-ca\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119759 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a1774b8-c928-4843-85ba-e02066fd0f5c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-jcsj5\" (UID: \"3a1774b8-c928-4843-85ba-e02066fd0f5c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.119805 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-oauth-serving-cert\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.120462 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-config\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.121155 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-client-ca\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.121678 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.121961 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.122145 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.122235 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.122245 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cd62560a-d47e-49f2-b3cd-06568dc86f43-audit-dir\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.122284 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.123000 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd62560a-d47e-49f2-b3cd-06568dc86f43-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.122140 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cd62560a-d47e-49f2-b3cd-06568dc86f43-audit-policies\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.124015 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.124904 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-images\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.125147 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-etcd-serving-ca\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.126253 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-audit-dir\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.150620 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.151728 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5fwj8"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.127219 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-image-import-ca\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.152227 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-encryption-config\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.152252 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-audit\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.152475 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cd62560a-d47e-49f2-b3cd-06568dc86f43-encryption-config\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.152472 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd62560a-d47e-49f2-b3cd-06568dc86f43-serving-cert\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.152593 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cd62560a-d47e-49f2-b3cd-06568dc86f43-etcd-client\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.152626 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-serving-cert\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.152874 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-node-pullsecrets\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.153144 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a2e9d7-a1ec-436d-8634-89c763644dec-serving-cert\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.153464 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-config\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.153545 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.153707 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cd62560a-d47e-49f2-b3cd-06568dc86f43-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.153732 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-config\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.153793 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.159765 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.159871 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.159782 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.160943 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tmk5l"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.161584 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.161611 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nj7lz"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.161623 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.161724 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.161734 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.162300 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4bvx"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.164074 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.164142 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.165210 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-etcd-client\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.165874 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-47kbw"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.167817 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nzjhz"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.171716 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.171761 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-6lcfr"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.174435 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5lb54"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.174671 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.176340 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.177175 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.180000 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.182453 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.184426 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.188033 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.188137 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.189619 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.193498 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zhxt6"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.194892 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-49gdq"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.196026 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-smwhg"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.196905 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.198019 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.199019 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.200083 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5w577"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.201133 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pq7h8"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.202135 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.203145 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-p8cxw"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.204125 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.205095 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5fwj8"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.206193 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-khj2k"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.207479 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-khj2k" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.207623 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-k785t"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.211690 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.212059 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.212085 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.213605 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.214529 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-khj2k"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.223883 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tmk5l"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224378 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a1774b8-c928-4843-85ba-e02066fd0f5c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-jcsj5\" (UID: \"3a1774b8-c928-4843-85ba-e02066fd0f5c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224428 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-oauth-serving-cert\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224482 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-client-ca\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224558 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-serving-cert\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224638 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-oauth-config\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224688 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6j99\" (UniqueName: \"kubernetes.io/projected/3a1774b8-c928-4843-85ba-e02066fd0f5c-kube-api-access-c6j99\") pod \"cluster-samples-operator-665b6dd947-jcsj5\" (UID: \"3a1774b8-c928-4843-85ba-e02066fd0f5c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224748 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31e65ba1-328b-4d4b-905c-6074ab5d1427-serving-cert\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224787 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8szn\" (UniqueName: \"kubernetes.io/projected/52e8e292-0788-4957-ac33-083fadb0f61b-kube-api-access-b8szn\") pod \"dns-operator-744455d44c-p8cxw\" (UID: \"52e8e292-0788-4957-ac33-083fadb0f61b\") " pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224813 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-config\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224889 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgf9m\" (UniqueName: \"kubernetes.io/projected/31e65ba1-328b-4d4b-905c-6074ab5d1427-kube-api-access-vgf9m\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224928 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-trusted-ca-bundle\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.224976 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c807324c-da31-4d36-8640-186d9cc31097-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.225028 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c807324c-da31-4d36-8640-186d9cc31097-metrics-tls\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.225068 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cbfn\" (UniqueName: \"kubernetes.io/projected/64fd38ac-d0ef-430c-aed5-5a4e4944986c-kube-api-access-2cbfn\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.225106 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-service-ca\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.225138 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/52e8e292-0788-4957-ac33-083fadb0f61b-metrics-tls\") pod \"dns-operator-744455d44c-p8cxw\" (UID: \"52e8e292-0788-4957-ac33-083fadb0f61b\") " pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.225188 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/31e65ba1-328b-4d4b-905c-6074ab5d1427-trusted-ca\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.225224 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/479ebf74-62ea-4516-8c29-8d9cdf128fb4-serving-cert\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.225259 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx9zz\" (UniqueName: \"kubernetes.io/projected/c807324c-da31-4d36-8640-186d9cc31097-kube-api-access-vx9zz\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.225319 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31e65ba1-328b-4d4b-905c-6074ab5d1427-config\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.225363 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wskjk\" (UniqueName: \"kubernetes.io/projected/479ebf74-62ea-4516-8c29-8d9cdf128fb4-kube-api-access-wskjk\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.225413 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c807324c-da31-4d36-8640-186d9cc31097-trusted-ca\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.225444 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-config\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.226066 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-client-ca\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.226148 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-oauth-serving-cert\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.226739 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-config\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.227113 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-config\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.227992 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.228176 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.228432 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31e65ba1-328b-4d4b-905c-6074ab5d1427-config\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.228951 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-trusted-ca-bundle\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.229620 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-service-ca\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.230203 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a1774b8-c928-4843-85ba-e02066fd0f5c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-jcsj5\" (UID: \"3a1774b8-c928-4843-85ba-e02066fd0f5c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.230877 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/31e65ba1-328b-4d4b-905c-6074ab5d1427-trusted-ca\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.231962 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.232073 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c807324c-da31-4d36-8640-186d9cc31097-trusted-ca\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.233508 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c807324c-da31-4d36-8640-186d9cc31097-metrics-tls\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.233715 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-serving-cert\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.233907 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31e65ba1-328b-4d4b-905c-6074ab5d1427-serving-cert\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.234212 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/479ebf74-62ea-4516-8c29-8d9cdf128fb4-serving-cert\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.235095 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.236014 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.237363 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-xxbtt"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.237574 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.238312 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xxbtt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.238928 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-k785t"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.240430 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6lcfr"] Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.243103 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-oauth-config\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.258081 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.284858 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.297995 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.317934 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.338472 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.351298 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/52e8e292-0788-4957-ac33-083fadb0f61b-metrics-tls\") pod \"dns-operator-744455d44c-p8cxw\" (UID: \"52e8e292-0788-4957-ac33-083fadb0f61b\") " pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.358187 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.377199 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.397251 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.418361 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.443635 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.458456 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.482299 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.498255 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.518349 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.538885 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.558315 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.570397 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.577752 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.619533 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.639843 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.658533 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.678112 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.698080 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.718555 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.738014 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.757522 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.797478 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.819496 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.837620 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.857735 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.877004 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.898359 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.917662 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.938137 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.958438 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.977167 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Sep 30 08:05:17 crc kubenswrapper[4810]: I0930 08:05:17.997832 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.018723 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.038810 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.055737 4810 request.go:700] Waited for 1.011247927s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.057953 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.078071 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.099220 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.118882 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.138488 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.160065 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.179508 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.197805 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.217929 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.239464 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.257716 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.278112 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.298142 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.316896 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.339556 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.357606 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.410395 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk7wv\" (UniqueName: \"kubernetes.io/projected/60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c-kube-api-access-mk7wv\") pod \"apiserver-76f77b778f-pq7h8\" (UID: \"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c\") " pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.428327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p895k\" (UniqueName: \"kubernetes.io/projected/f7201c6b-de6b-4c4e-b20b-90c59d0db7ac-kube-api-access-p895k\") pod \"machine-api-operator-5694c8668f-nj7lz\" (UID: \"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.451388 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqp5h\" (UniqueName: \"kubernetes.io/projected/cd62560a-d47e-49f2-b3cd-06568dc86f43-kube-api-access-rqp5h\") pod \"apiserver-7bbb656c7d-gnt7r\" (UID: \"cd62560a-d47e-49f2-b3cd-06568dc86f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.457677 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.467843 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzj7l\" (UniqueName: \"kubernetes.io/projected/80a2e9d7-a1ec-436d-8634-89c763644dec-kube-api-access-gzj7l\") pod \"controller-manager-879f6c89f-qdcml\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.478420 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.498056 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.513178 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.518034 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.525743 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.538322 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.558789 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.577486 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.598459 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.618558 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.640876 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.662474 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.678167 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.697106 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.723829 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.728478 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.742673 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.743064 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.758160 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.778560 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.798448 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.818201 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.843374 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.856614 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pq7h8"] Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.858802 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Sep 30 08:05:18 crc kubenswrapper[4810]: W0930 08:05:18.869675 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60f55bdb_8ad2_4e65_b3c2_aed4488e6a1c.slice/crio-2b91636a80ffb37b872919c63063ed8f5958909c2bc93e9609751b456d4d7a06 WatchSource:0}: Error finding container 2b91636a80ffb37b872919c63063ed8f5958909c2bc93e9609751b456d4d7a06: Status 404 returned error can't find the container with id 2b91636a80ffb37b872919c63063ed8f5958909c2bc93e9609751b456d4d7a06 Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.877480 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.885605 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nj7lz"] Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.900846 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.917718 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.937503 4810 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.960174 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r"] Sep 30 08:05:18 crc kubenswrapper[4810]: W0930 08:05:18.971898 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd62560a_d47e_49f2_b3cd_06568dc86f43.slice/crio-b79efa66a041ddfc033fe30d8fd4e785df17a52edd5febd43861371447b959ea WatchSource:0}: Error finding container b79efa66a041ddfc033fe30d8fd4e785df17a52edd5febd43861371447b959ea: Status 404 returned error can't find the container with id b79efa66a041ddfc033fe30d8fd4e785df17a52edd5febd43861371447b959ea Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.988295 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8szn\" (UniqueName: \"kubernetes.io/projected/52e8e292-0788-4957-ac33-083fadb0f61b-kube-api-access-b8szn\") pod \"dns-operator-744455d44c-p8cxw\" (UID: \"52e8e292-0788-4957-ac33-083fadb0f61b\") " pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.998367 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qdcml"] Sep 30 08:05:18 crc kubenswrapper[4810]: I0930 08:05:18.998822 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgf9m\" (UniqueName: \"kubernetes.io/projected/31e65ba1-328b-4d4b-905c-6074ab5d1427-kube-api-access-vgf9m\") pod \"console-operator-58897d9998-49gdq\" (UID: \"31e65ba1-328b-4d4b-905c-6074ab5d1427\") " pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.012059 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.016461 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6j99\" (UniqueName: \"kubernetes.io/projected/3a1774b8-c928-4843-85ba-e02066fd0f5c-kube-api-access-c6j99\") pod \"cluster-samples-operator-665b6dd947-jcsj5\" (UID: \"3a1774b8-c928-4843-85ba-e02066fd0f5c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.038352 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cbfn\" (UniqueName: \"kubernetes.io/projected/64fd38ac-d0ef-430c-aed5-5a4e4944986c-kube-api-access-2cbfn\") pod \"console-f9d7485db-5w577\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.052482 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c807324c-da31-4d36-8640-186d9cc31097-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.055156 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.055970 4810 request.go:700] Waited for 1.826703669s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/serviceaccounts/route-controller-manager-sa/token Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.072734 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wskjk\" (UniqueName: \"kubernetes.io/projected/479ebf74-62ea-4516-8c29-8d9cdf128fb4-kube-api-access-wskjk\") pod \"route-controller-manager-6576b87f9c-xggpf\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.084007 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.094884 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx9zz\" (UniqueName: \"kubernetes.io/projected/c807324c-da31-4d36-8640-186d9cc31097-kube-api-access-vx9zz\") pod \"ingress-operator-5b745b69d9-mnhwj\" (UID: \"c807324c-da31-4d36-8640-186d9cc31097\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.096924 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.117851 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.138056 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.139125 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.194157 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" event={"ID":"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c","Type":"ContainerStarted","Data":"2b91636a80ffb37b872919c63063ed8f5958909c2bc93e9609751b456d4d7a06"} Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.208449 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" event={"ID":"cd62560a-d47e-49f2-b3cd-06568dc86f43","Type":"ContainerStarted","Data":"b79efa66a041ddfc033fe30d8fd4e785df17a52edd5febd43861371447b959ea"} Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.213793 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" event={"ID":"80a2e9d7-a1ec-436d-8634-89c763644dec","Type":"ContainerStarted","Data":"21c1fbdeff2ead362912ec455538dc597a1558b6220b8195e0267d21dab955b3"} Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.214867 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" event={"ID":"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac","Type":"ContainerStarted","Data":"30c3bdeed937a823ed9dd3a82c0792d0a859c51d21e019e9afd58d6276edf52e"} Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.214898 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" event={"ID":"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac","Type":"ContainerStarted","Data":"fbefb92615002791bc1472354a3fc350cd4c25fdd990ee5c8d5fcc4aba93cae7"} Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.247380 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252568 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k965l\" (UniqueName: \"kubernetes.io/projected/626d8738-cb60-49b6-ba04-754600d28a86-kube-api-access-k965l\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252612 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252654 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-bound-sa-token\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252671 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc4cz\" (UniqueName: \"kubernetes.io/projected/95f2618f-a0a5-43a2-b639-25944e913395-kube-api-access-zc4cz\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252688 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4rzw\" (UniqueName: \"kubernetes.io/projected/f4a58c1d-0d71-4853-b894-e14efec97e65-kube-api-access-w4rzw\") pod \"openshift-config-operator-7777fb866f-rg6tm\" (UID: \"f4a58c1d-0d71-4853-b894-e14efec97e65\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252705 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmjfj\" (UniqueName: \"kubernetes.io/projected/e57275b7-e952-4962-89fd-7950616d05f0-kube-api-access-lmjfj\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252720 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-certificates\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252738 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkl2r\" (UniqueName: \"kubernetes.io/projected/f840f471-2128-47ce-9bd6-10eb24c31be5-kube-api-access-tkl2r\") pod \"openshift-controller-manager-operator-756b6f6bc6-5wczx\" (UID: \"f840f471-2128-47ce-9bd6-10eb24c31be5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252752 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n97fz\" (UniqueName: \"kubernetes.io/projected/e90bfa19-6003-4593-b318-31ab8ec8c44f-kube-api-access-n97fz\") pod \"downloads-7954f5f757-47kbw\" (UID: \"e90bfa19-6003-4593-b318-31ab8ec8c44f\") " pod="openshift-console/downloads-7954f5f757-47kbw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252793 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252809 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/626d8738-cb60-49b6-ba04-754600d28a86-machine-approver-tls\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252825 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e57275b7-e952-4962-89fd-7950616d05f0-serving-cert\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252851 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f4a58c1d-0d71-4853-b894-e14efec97e65-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rg6tm\" (UID: \"f4a58c1d-0d71-4853-b894-e14efec97e65\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252866 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252883 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/626d8738-cb60-49b6-ba04-754600d28a86-auth-proxy-config\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.252970 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4c2e4c71-7e20-4d81-8531-f673f4c9341f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253005 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-serving-cert\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253045 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e57275b7-e952-4962-89fd-7950616d05f0-config\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253066 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253088 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-config\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253111 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6083d85f-fd9a-4df3-84c5-f6a92e84c53c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-49zt7\" (UID: \"6083d85f-fd9a-4df3-84c5-f6a92e84c53c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253139 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6083d85f-fd9a-4df3-84c5-f6a92e84c53c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-49zt7\" (UID: \"6083d85f-fd9a-4df3-84c5-f6a92e84c53c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253155 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d60f7e34-64f7-4364-b1a3-de505ad145df-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: E0930 08:05:19.253178 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:19.753164081 +0000 UTC m=+143.205363338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253316 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d60f7e34-64f7-4364-b1a3-de505ad145df-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253354 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-service-ca-bundle\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253373 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253411 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e57275b7-e952-4962-89fd-7950616d05f0-etcd-ca\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253432 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cc78ba14-6088-4423-932c-9377a1c99a2d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fkmsl\" (UID: \"cc78ba14-6088-4423-932c-9377a1c99a2d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253493 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f840f471-2128-47ce-9bd6-10eb24c31be5-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5wczx\" (UID: \"f840f471-2128-47ce-9bd6-10eb24c31be5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253510 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6v92\" (UniqueName: \"kubernetes.io/projected/5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a-kube-api-access-q6v92\") pod \"openshift-apiserver-operator-796bbdcf4f-s2vww\" (UID: \"5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253528 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253546 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-trusted-ca\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253677 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4c2e4c71-7e20-4d81-8531-f673f4c9341f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253696 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f840f471-2128-47ce-9bd6-10eb24c31be5-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5wczx\" (UID: \"f840f471-2128-47ce-9bd6-10eb24c31be5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253713 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4a58c1d-0d71-4853-b894-e14efec97e65-serving-cert\") pod \"openshift-config-operator-7777fb866f-rg6tm\" (UID: \"f4a58c1d-0d71-4853-b894-e14efec97e65\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253734 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-s2vww\" (UID: \"5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253756 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-tls\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253800 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs9h6\" (UniqueName: \"kubernetes.io/projected/6760d07c-3b94-492f-b58c-c61dbd1df867-kube-api-access-gs9h6\") pod \"multus-admission-controller-857f4d67dd-5lb54\" (UID: \"6760d07c-3b94-492f-b58c-c61dbd1df867\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253817 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253837 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253877 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253918 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8sgm\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-kube-api-access-n8sgm\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253956 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-audit-policies\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.253971 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.255160 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.255311 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt7c7\" (UniqueName: \"kubernetes.io/projected/6083d85f-fd9a-4df3-84c5-f6a92e84c53c-kube-api-access-rt7c7\") pod \"kube-storage-version-migrator-operator-b67b599dd-49zt7\" (UID: \"6083d85f-fd9a-4df3-84c5-f6a92e84c53c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.255389 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-s2vww\" (UID: \"5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.255448 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d60f7e34-64f7-4364-b1a3-de505ad145df-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.255483 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj8x6\" (UniqueName: \"kubernetes.io/projected/d60f7e34-64f7-4364-b1a3-de505ad145df-kube-api-access-jj8x6\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.255540 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e57275b7-e952-4962-89fd-7950616d05f0-etcd-service-ca\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.255836 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.255961 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/626d8738-cb60-49b6-ba04-754600d28a86-config\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.255985 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.256002 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6760d07c-3b94-492f-b58c-c61dbd1df867-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5lb54\" (UID: \"6760d07c-3b94-492f-b58c-c61dbd1df867\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.256039 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/95f2618f-a0a5-43a2-b639-25944e913395-audit-dir\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.256057 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc78ba14-6088-4423-932c-9377a1c99a2d-config\") pod \"kube-controller-manager-operator-78b949d7b-fkmsl\" (UID: \"cc78ba14-6088-4423-932c-9377a1c99a2d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.256087 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxxj2\" (UniqueName: \"kubernetes.io/projected/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-kube-api-access-wxxj2\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.256122 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e57275b7-e952-4962-89fd-7950616d05f0-etcd-client\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.256137 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc78ba14-6088-4423-932c-9377a1c99a2d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fkmsl\" (UID: \"cc78ba14-6088-4423-932c-9377a1c99a2d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.285496 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-49gdq"] Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.356768 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:19 crc kubenswrapper[4810]: E0930 08:05:19.356942 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:19.856909984 +0000 UTC m=+143.309109241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.357423 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.357488 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.358153 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.358197 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8sgm\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-kube-api-access-n8sgm\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.358427 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.358523 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.358555 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/689650fc-6a12-4737-987d-1b1205e46e41-profile-collector-cert\") pod \"catalog-operator-68c6474976-rfd6l\" (UID: \"689650fc-6a12-4737-987d-1b1205e46e41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.358580 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-config-volume\") pod \"collect-profiles-29320320-kzpbn\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.358608 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sck96\" (UniqueName: \"kubernetes.io/projected/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-kube-api-access-sck96\") pod \"collect-profiles-29320320-kzpbn\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.358755 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/626d8738-cb60-49b6-ba04-754600d28a86-config\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.358932 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6760d07c-3b94-492f-b58c-c61dbd1df867-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5lb54\" (UID: \"6760d07c-3b94-492f-b58c-c61dbd1df867\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.359035 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/95f2618f-a0a5-43a2-b639-25944e913395-audit-dir\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.359163 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f8d81122-beac-495b-9e7f-76fad0fd5fa1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zzks6\" (UID: \"f8d81122-beac-495b-9e7f-76fad0fd5fa1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.359220 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/95f2618f-a0a5-43a2-b639-25944e913395-audit-dir\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.359255 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/626d8738-cb60-49b6-ba04-754600d28a86-config\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.359290 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e57275b7-e952-4962-89fd-7950616d05f0-etcd-client\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.360128 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc78ba14-6088-4423-932c-9377a1c99a2d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fkmsl\" (UID: \"cc78ba14-6088-4423-932c-9377a1c99a2d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.363674 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.363696 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e57275b7-e952-4962-89fd-7950616d05f0-etcd-client\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.363736 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-mountpoint-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.363821 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6r2j\" (UniqueName: \"kubernetes.io/projected/00c025fa-c660-4977-a0f6-bf3d1c64ccd2-kube-api-access-t6r2j\") pod \"service-ca-9c57cc56f-5fwj8\" (UID: \"00c025fa-c660-4977-a0f6-bf3d1c64ccd2\") " pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.363851 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5df805e2-6930-4217-adbb-8c1344b15fbb-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wfvkx\" (UID: \"5df805e2-6930-4217-adbb-8c1344b15fbb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.364708 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6760d07c-3b94-492f-b58c-c61dbd1df867-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5lb54\" (UID: \"6760d07c-3b94-492f-b58c-c61dbd1df867\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.365336 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7qb9\" (UniqueName: \"kubernetes.io/projected/1c38bb35-d8db-40f6-903a-10682aa99f35-kube-api-access-v7qb9\") pod \"dns-default-6lcfr\" (UID: \"1c38bb35-d8db-40f6-903a-10682aa99f35\") " pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.365430 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4-serving-cert\") pod \"service-ca-operator-777779d784-4mdfd\" (UID: \"fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.365500 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-bound-sa-token\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.365561 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4rzw\" (UniqueName: \"kubernetes.io/projected/f4a58c1d-0d71-4853-b894-e14efec97e65-kube-api-access-w4rzw\") pod \"openshift-config-operator-7777fb866f-rg6tm\" (UID: \"f4a58c1d-0d71-4853-b894-e14efec97e65\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.365607 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06d3ea99-b8e6-4e46-bd04-e370f73274ef-config\") pod \"kube-apiserver-operator-766d6c64bb-6gc9d\" (UID: \"06d3ea99-b8e6-4e46-bd04-e370f73274ef\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.365665 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-certificates\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.365692 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n97fz\" (UniqueName: \"kubernetes.io/projected/e90bfa19-6003-4593-b318-31ab8ec8c44f-kube-api-access-n97fz\") pod \"downloads-7954f5f757-47kbw\" (UID: \"e90bfa19-6003-4593-b318-31ab8ec8c44f\") " pod="openshift-console/downloads-7954f5f757-47kbw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366120 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc78ba14-6088-4423-932c-9377a1c99a2d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fkmsl\" (UID: \"cc78ba14-6088-4423-932c-9377a1c99a2d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366230 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a7250b9a-5aef-438b-b767-430287477d1e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5h6v8\" (UID: \"a7250b9a-5aef-438b-b767-430287477d1e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366318 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a7250b9a-5aef-438b-b767-430287477d1e-srv-cert\") pod \"olm-operator-6b444d44fb-5h6v8\" (UID: \"a7250b9a-5aef-438b-b767-430287477d1e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366345 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brtxg\" (UniqueName: \"kubernetes.io/projected/167864ca-6a75-4fc0-961a-9af57dcce611-kube-api-access-brtxg\") pod \"ingress-canary-khj2k\" (UID: \"167864ca-6a75-4fc0-961a-9af57dcce611\") " pod="openshift-ingress-canary/ingress-canary-khj2k" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366414 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366488 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-tmpfs\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366594 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f4b41c70-05a1-4342-8aee-cba6e44b89dc-metrics-certs\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366702 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4c2e4c71-7e20-4d81-8531-f673f4c9341f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366758 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-apiservice-cert\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366784 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f4b41c70-05a1-4342-8aee-cba6e44b89dc-default-certificate\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366846 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e57275b7-e952-4962-89fd-7950616d05f0-config\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.366880 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-proxy-tls\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.367111 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6083d85f-fd9a-4df3-84c5-f6a92e84c53c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-49zt7\" (UID: \"6083d85f-fd9a-4df3-84c5-f6a92e84c53c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.367136 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1c38bb35-d8db-40f6-903a-10682aa99f35-metrics-tls\") pod \"dns-default-6lcfr\" (UID: \"1c38bb35-d8db-40f6-903a-10682aa99f35\") " pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.367157 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5892\" (UniqueName: \"kubernetes.io/projected/e049bda6-e38f-4357-926e-d8b767cf5ed9-kube-api-access-m5892\") pod \"marketplace-operator-79b997595-tmk5l\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.367197 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/00e1ebdd-3664-401d-b7bf-a9f786d9b6d2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p7cc4\" (UID: \"00e1ebdd-3664-401d-b7bf-a9f786d9b6d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.367224 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6083d85f-fd9a-4df3-84c5-f6a92e84c53c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-49zt7\" (UID: \"6083d85f-fd9a-4df3-84c5-f6a92e84c53c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.367244 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d60f7e34-64f7-4364-b1a3-de505ad145df-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.367309 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1c38bb35-d8db-40f6-903a-10682aa99f35-config-volume\") pod \"dns-default-6lcfr\" (UID: \"1c38bb35-d8db-40f6-903a-10682aa99f35\") " pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.367623 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/00c025fa-c660-4977-a0f6-bf3d1c64ccd2-signing-key\") pod \"service-ca-9c57cc56f-5fwj8\" (UID: \"00c025fa-c660-4977-a0f6-bf3d1c64ccd2\") " pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.367654 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.367813 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5971693-8b44-4a2a-bd8f-7ee41e0f69cb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-blbjc\" (UID: \"c5971693-8b44-4a2a-bd8f-7ee41e0f69cb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.367628 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-certificates\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.368285 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.368666 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e57275b7-e952-4962-89fd-7950616d05f0-config\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.368672 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tmk5l\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.368722 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f4b41c70-05a1-4342-8aee-cba6e44b89dc-stats-auth\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.368748 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-service-ca-bundle\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.368787 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d60f7e34-64f7-4364-b1a3-de505ad145df-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370003 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cc78ba14-6088-4423-932c-9377a1c99a2d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fkmsl\" (UID: \"cc78ba14-6088-4423-932c-9377a1c99a2d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370085 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e57275b7-e952-4962-89fd-7950616d05f0-etcd-ca\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370142 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6v92\" (UniqueName: \"kubernetes.io/projected/5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a-kube-api-access-q6v92\") pod \"openshift-apiserver-operator-796bbdcf4f-s2vww\" (UID: \"5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370154 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d60f7e34-64f7-4364-b1a3-de505ad145df-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370168 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-socket-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.369180 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6083d85f-fd9a-4df3-84c5-f6a92e84c53c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-49zt7\" (UID: \"6083d85f-fd9a-4df3-84c5-f6a92e84c53c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.369550 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370198 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tmk5l\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370219 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4a58c1d-0d71-4853-b894-e14efec97e65-serving-cert\") pod \"openshift-config-operator-7777fb866f-rg6tm\" (UID: \"f4a58c1d-0d71-4853-b894-e14efec97e65\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370243 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/689650fc-6a12-4737-987d-1b1205e46e41-srv-cert\") pod \"catalog-operator-68c6474976-rfd6l\" (UID: \"689650fc-6a12-4737-987d-1b1205e46e41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370277 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-tls\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370293 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs9h6\" (UniqueName: \"kubernetes.io/projected/6760d07c-3b94-492f-b58c-c61dbd1df867-kube-api-access-gs9h6\") pod \"multus-admission-controller-857f4d67dd-5lb54\" (UID: \"6760d07c-3b94-492f-b58c-c61dbd1df867\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370315 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnvvq\" (UniqueName: \"kubernetes.io/projected/a7250b9a-5aef-438b-b767-430287477d1e-kube-api-access-hnvvq\") pod \"olm-operator-6b444d44fb-5h6v8\" (UID: \"a7250b9a-5aef-438b-b767-430287477d1e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370333 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rpx5\" (UniqueName: \"kubernetes.io/projected/00e1ebdd-3664-401d-b7bf-a9f786d9b6d2-kube-api-access-5rpx5\") pod \"control-plane-machine-set-operator-78cbb6b69f-p7cc4\" (UID: \"00e1ebdd-3664-401d-b7bf-a9f786d9b6d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370348 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/167864ca-6a75-4fc0-961a-9af57dcce611-cert\") pod \"ingress-canary-khj2k\" (UID: \"167864ca-6a75-4fc0-961a-9af57dcce611\") " pod="openshift-ingress-canary/ingress-canary-khj2k" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370363 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv99d\" (UniqueName: \"kubernetes.io/projected/f4b41c70-05a1-4342-8aee-cba6e44b89dc-kube-api-access-gv99d\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370402 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-audit-policies\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370418 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vdsr\" (UniqueName: \"kubernetes.io/projected/5b361c60-8056-4c29-a634-97c8a1c58f67-kube-api-access-8vdsr\") pod \"machine-config-server-xxbtt\" (UID: \"5b361c60-8056-4c29-a634-97c8a1c58f67\") " pod="openshift-machine-config-operator/machine-config-server-xxbtt" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370435 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4b41c70-05a1-4342-8aee-cba6e44b89dc-service-ca-bundle\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370464 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-s2vww\" (UID: \"5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370481 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d60f7e34-64f7-4364-b1a3-de505ad145df-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370496 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj8x6\" (UniqueName: \"kubernetes.io/projected/d60f7e34-64f7-4364-b1a3-de505ad145df-kube-api-access-jj8x6\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370511 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e57275b7-e952-4962-89fd-7950616d05f0-etcd-service-ca\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370530 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl8vd\" (UniqueName: \"kubernetes.io/projected/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-kube-api-access-bl8vd\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370548 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt7c7\" (UniqueName: \"kubernetes.io/projected/6083d85f-fd9a-4df3-84c5-f6a92e84c53c-kube-api-access-rt7c7\") pod \"kube-storage-version-migrator-operator-b67b599dd-49zt7\" (UID: \"6083d85f-fd9a-4df3-84c5-f6a92e84c53c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370566 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc78ba14-6088-4423-932c-9377a1c99a2d-config\") pod \"kube-controller-manager-operator-78b949d7b-fkmsl\" (UID: \"cc78ba14-6088-4423-932c-9377a1c99a2d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370607 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/00c025fa-c660-4977-a0f6-bf3d1c64ccd2-signing-cabundle\") pod \"service-ca-9c57cc56f-5fwj8\" (UID: \"00c025fa-c660-4977-a0f6-bf3d1c64ccd2\") " pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370627 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4sbv\" (UniqueName: \"kubernetes.io/projected/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-kube-api-access-z4sbv\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370645 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8d81122-beac-495b-9e7f-76fad0fd5fa1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zzks6\" (UID: \"f8d81122-beac-495b-9e7f-76fad0fd5fa1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370663 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370669 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e57275b7-e952-4962-89fd-7950616d05f0-etcd-ca\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370680 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.369200 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4c2e4c71-7e20-4d81-8531-f673f4c9341f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.369810 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370698 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxxj2\" (UniqueName: \"kubernetes.io/projected/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-kube-api-access-wxxj2\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.369918 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-service-ca-bundle\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.370773 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-registration-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.371229 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-csi-data-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.371750 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6083d85f-fd9a-4df3-84c5-f6a92e84c53c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-49zt7\" (UID: \"6083d85f-fd9a-4df3-84c5-f6a92e84c53c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.371908 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d60f7e34-64f7-4364-b1a3-de505ad145df-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.371259 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p22n\" (UniqueName: \"kubernetes.io/projected/5df805e2-6930-4217-adbb-8c1344b15fbb-kube-api-access-4p22n\") pod \"machine-config-controller-84d6567774-wfvkx\" (UID: \"5df805e2-6930-4217-adbb-8c1344b15fbb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.371972 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsqlm\" (UniqueName: \"kubernetes.io/projected/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-kube-api-access-tsqlm\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372082 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k965l\" (UniqueName: \"kubernetes.io/projected/626d8738-cb60-49b6-ba04-754600d28a86-kube-api-access-k965l\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372124 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlfxs\" (UniqueName: \"kubernetes.io/projected/fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4-kube-api-access-tlfxs\") pod \"service-ca-operator-777779d784-4mdfd\" (UID: \"fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372329 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4-config\") pod \"service-ca-operator-777779d784-4mdfd\" (UID: \"fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372389 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372410 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc4cz\" (UniqueName: \"kubernetes.io/projected/95f2618f-a0a5-43a2-b639-25944e913395-kube-api-access-zc4cz\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372459 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-audit-policies\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372602 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmjfj\" (UniqueName: \"kubernetes.io/projected/e57275b7-e952-4962-89fd-7950616d05f0-kube-api-access-lmjfj\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372642 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp8nw\" (UniqueName: \"kubernetes.io/projected/c5971693-8b44-4a2a-bd8f-7ee41e0f69cb-kube-api-access-bp8nw\") pod \"package-server-manager-789f6589d5-blbjc\" (UID: \"c5971693-8b44-4a2a-bd8f-7ee41e0f69cb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372724 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkl2r\" (UniqueName: \"kubernetes.io/projected/f840f471-2128-47ce-9bd6-10eb24c31be5-kube-api-access-tkl2r\") pod \"openshift-controller-manager-operator-756b6f6bc6-5wczx\" (UID: \"f840f471-2128-47ce-9bd6-10eb24c31be5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372801 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372855 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/626d8738-cb60-49b6-ba04-754600d28a86-machine-approver-tls\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372880 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e57275b7-e952-4962-89fd-7950616d05f0-serving-cert\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372902 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-plugins-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372920 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbdjr\" (UniqueName: \"kubernetes.io/projected/bc765a10-3756-403a-8723-2cb47e961f26-kube-api-access-hbdjr\") pod \"migrator-59844c95c7-xtbw5\" (UID: \"bc765a10-3756-403a-8723-2cb47e961f26\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.372938 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06d3ea99-b8e6-4e46-bd04-e370f73274ef-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6gc9d\" (UID: \"06d3ea99-b8e6-4e46-bd04-e370f73274ef\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" Sep 30 08:05:19 crc kubenswrapper[4810]: E0930 08:05:19.373111 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:19.873097802 +0000 UTC m=+143.325297059 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.373297 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.373433 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-secret-volume\") pod \"collect-profiles-29320320-kzpbn\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.373487 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-images\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.373586 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f4a58c1d-0d71-4853-b894-e14efec97e65-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rg6tm\" (UID: \"f4a58c1d-0d71-4853-b894-e14efec97e65\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.373634 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4a58c1d-0d71-4853-b894-e14efec97e65-serving-cert\") pod \"openshift-config-operator-7777fb866f-rg6tm\" (UID: \"f4a58c1d-0d71-4853-b894-e14efec97e65\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.373650 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/626d8738-cb60-49b6-ba04-754600d28a86-auth-proxy-config\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.373681 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5b361c60-8056-4c29-a634-97c8a1c58f67-certs\") pod \"machine-config-server-xxbtt\" (UID: \"5b361c60-8056-4c29-a634-97c8a1c58f67\") " pod="openshift-machine-config-operator/machine-config-server-xxbtt" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.373706 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06d3ea99-b8e6-4e46-bd04-e370f73274ef-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6gc9d\" (UID: \"06d3ea99-b8e6-4e46-bd04-e370f73274ef\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.373820 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-serving-cert\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.373841 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-webhook-cert\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374107 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/626d8738-cb60-49b6-ba04-754600d28a86-auth-proxy-config\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374138 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-tls\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374180 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374227 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-config\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374248 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctkkz\" (UniqueName: \"kubernetes.io/projected/689650fc-6a12-4737-987d-1b1205e46e41-kube-api-access-ctkkz\") pod \"catalog-operator-68c6474976-rfd6l\" (UID: \"689650fc-6a12-4737-987d-1b1205e46e41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374287 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5df805e2-6930-4217-adbb-8c1344b15fbb-proxy-tls\") pod \"machine-config-controller-84d6567774-wfvkx\" (UID: \"5df805e2-6930-4217-adbb-8c1344b15fbb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374477 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374678 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f840f471-2128-47ce-9bd6-10eb24c31be5-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5wczx\" (UID: \"f840f471-2128-47ce-9bd6-10eb24c31be5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374720 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374747 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8d81122-beac-495b-9e7f-76fad0fd5fa1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zzks6\" (UID: \"f8d81122-beac-495b-9e7f-76fad0fd5fa1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374765 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-trusted-ca\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374812 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4c2e4c71-7e20-4d81-8531-f673f4c9341f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374832 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f840f471-2128-47ce-9bd6-10eb24c31be5-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5wczx\" (UID: \"f840f471-2128-47ce-9bd6-10eb24c31be5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374853 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-s2vww\" (UID: \"5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.374872 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5b361c60-8056-4c29-a634-97c8a1c58f67-node-bootstrap-token\") pod \"machine-config-server-xxbtt\" (UID: \"5b361c60-8056-4c29-a634-97c8a1c58f67\") " pod="openshift-machine-config-operator/machine-config-server-xxbtt" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.375076 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f4a58c1d-0d71-4853-b894-e14efec97e65-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rg6tm\" (UID: \"f4a58c1d-0d71-4853-b894-e14efec97e65\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.375570 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-config\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.377839 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc78ba14-6088-4423-932c-9377a1c99a2d-config\") pod \"kube-controller-manager-operator-78b949d7b-fkmsl\" (UID: \"cc78ba14-6088-4423-932c-9377a1c99a2d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.378759 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-trusted-ca\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.379989 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.379998 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e57275b7-e952-4962-89fd-7950616d05f0-serving-cert\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.380341 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-s2vww\" (UID: \"5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.380470 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.380859 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-s2vww\" (UID: \"5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.380888 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f840f471-2128-47ce-9bd6-10eb24c31be5-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5wczx\" (UID: \"f840f471-2128-47ce-9bd6-10eb24c31be5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.380955 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.376225 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e57275b7-e952-4962-89fd-7950616d05f0-etcd-service-ca\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.383883 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4c2e4c71-7e20-4d81-8531-f673f4c9341f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.384704 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f840f471-2128-47ce-9bd6-10eb24c31be5-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5wczx\" (UID: \"f840f471-2128-47ce-9bd6-10eb24c31be5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.385391 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.385538 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-serving-cert\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.387130 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/626d8738-cb60-49b6-ba04-754600d28a86-machine-approver-tls\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.387353 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.387509 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5"] Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.391323 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-p8cxw"] Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.394580 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.395394 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.414991 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8sgm\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-kube-api-access-n8sgm\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.435086 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4rzw\" (UniqueName: \"kubernetes.io/projected/f4a58c1d-0d71-4853-b894-e14efec97e65-kube-api-access-w4rzw\") pod \"openshift-config-operator-7777fb866f-rg6tm\" (UID: \"f4a58c1d-0d71-4853-b894-e14efec97e65\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.454510 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-bound-sa-token\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.474108 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n97fz\" (UniqueName: \"kubernetes.io/projected/e90bfa19-6003-4593-b318-31ab8ec8c44f-kube-api-access-n97fz\") pod \"downloads-7954f5f757-47kbw\" (UID: \"e90bfa19-6003-4593-b318-31ab8ec8c44f\") " pod="openshift-console/downloads-7954f5f757-47kbw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.485986 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:19 crc kubenswrapper[4810]: E0930 08:05:19.486172 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:19.986146488 +0000 UTC m=+143.438345765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486432 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5892\" (UniqueName: \"kubernetes.io/projected/e049bda6-e38f-4357-926e-d8b767cf5ed9-kube-api-access-m5892\") pod \"marketplace-operator-79b997595-tmk5l\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/00e1ebdd-3664-401d-b7bf-a9f786d9b6d2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p7cc4\" (UID: \"00e1ebdd-3664-401d-b7bf-a9f786d9b6d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486493 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1c38bb35-d8db-40f6-903a-10682aa99f35-config-volume\") pod \"dns-default-6lcfr\" (UID: \"1c38bb35-d8db-40f6-903a-10682aa99f35\") " pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486530 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tmk5l\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486551 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f4b41c70-05a1-4342-8aee-cba6e44b89dc-stats-auth\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486571 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/00c025fa-c660-4977-a0f6-bf3d1c64ccd2-signing-key\") pod \"service-ca-9c57cc56f-5fwj8\" (UID: \"00c025fa-c660-4977-a0f6-bf3d1c64ccd2\") " pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5971693-8b44-4a2a-bd8f-7ee41e0f69cb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-blbjc\" (UID: \"c5971693-8b44-4a2a-bd8f-7ee41e0f69cb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486627 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-socket-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486654 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tmk5l\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486676 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/689650fc-6a12-4737-987d-1b1205e46e41-srv-cert\") pod \"catalog-operator-68c6474976-rfd6l\" (UID: \"689650fc-6a12-4737-987d-1b1205e46e41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486698 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnvvq\" (UniqueName: \"kubernetes.io/projected/a7250b9a-5aef-438b-b767-430287477d1e-kube-api-access-hnvvq\") pod \"olm-operator-6b444d44fb-5h6v8\" (UID: \"a7250b9a-5aef-438b-b767-430287477d1e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486722 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rpx5\" (UniqueName: \"kubernetes.io/projected/00e1ebdd-3664-401d-b7bf-a9f786d9b6d2-kube-api-access-5rpx5\") pod \"control-plane-machine-set-operator-78cbb6b69f-p7cc4\" (UID: \"00e1ebdd-3664-401d-b7bf-a9f786d9b6d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486744 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/167864ca-6a75-4fc0-961a-9af57dcce611-cert\") pod \"ingress-canary-khj2k\" (UID: \"167864ca-6a75-4fc0-961a-9af57dcce611\") " pod="openshift-ingress-canary/ingress-canary-khj2k" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486770 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv99d\" (UniqueName: \"kubernetes.io/projected/f4b41c70-05a1-4342-8aee-cba6e44b89dc-kube-api-access-gv99d\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486811 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4b41c70-05a1-4342-8aee-cba6e44b89dc-service-ca-bundle\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486835 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vdsr\" (UniqueName: \"kubernetes.io/projected/5b361c60-8056-4c29-a634-97c8a1c58f67-kube-api-access-8vdsr\") pod \"machine-config-server-xxbtt\" (UID: \"5b361c60-8056-4c29-a634-97c8a1c58f67\") " pod="openshift-machine-config-operator/machine-config-server-xxbtt" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486870 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl8vd\" (UniqueName: \"kubernetes.io/projected/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-kube-api-access-bl8vd\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486913 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4sbv\" (UniqueName: \"kubernetes.io/projected/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-kube-api-access-z4sbv\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486938 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8d81122-beac-495b-9e7f-76fad0fd5fa1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zzks6\" (UID: \"f8d81122-beac-495b-9e7f-76fad0fd5fa1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.486961 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487008 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/00c025fa-c660-4977-a0f6-bf3d1c64ccd2-signing-cabundle\") pod \"service-ca-9c57cc56f-5fwj8\" (UID: \"00c025fa-c660-4977-a0f6-bf3d1c64ccd2\") " pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487034 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-registration-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487059 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-csi-data-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487084 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p22n\" (UniqueName: \"kubernetes.io/projected/5df805e2-6930-4217-adbb-8c1344b15fbb-kube-api-access-4p22n\") pod \"machine-config-controller-84d6567774-wfvkx\" (UID: \"5df805e2-6930-4217-adbb-8c1344b15fbb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487117 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsqlm\" (UniqueName: \"kubernetes.io/projected/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-kube-api-access-tsqlm\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487140 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlfxs\" (UniqueName: \"kubernetes.io/projected/fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4-kube-api-access-tlfxs\") pod \"service-ca-operator-777779d784-4mdfd\" (UID: \"fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487173 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4-config\") pod \"service-ca-operator-777779d784-4mdfd\" (UID: \"fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487222 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp8nw\" (UniqueName: \"kubernetes.io/projected/c5971693-8b44-4a2a-bd8f-7ee41e0f69cb-kube-api-access-bp8nw\") pod \"package-server-manager-789f6589d5-blbjc\" (UID: \"c5971693-8b44-4a2a-bd8f-7ee41e0f69cb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487279 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-plugins-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487301 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbdjr\" (UniqueName: \"kubernetes.io/projected/bc765a10-3756-403a-8723-2cb47e961f26-kube-api-access-hbdjr\") pod \"migrator-59844c95c7-xtbw5\" (UID: \"bc765a10-3756-403a-8723-2cb47e961f26\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487324 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06d3ea99-b8e6-4e46-bd04-e370f73274ef-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6gc9d\" (UID: \"06d3ea99-b8e6-4e46-bd04-e370f73274ef\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487350 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487378 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-secret-volume\") pod \"collect-profiles-29320320-kzpbn\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487402 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-images\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487428 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06d3ea99-b8e6-4e46-bd04-e370f73274ef-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6gc9d\" (UID: \"06d3ea99-b8e6-4e46-bd04-e370f73274ef\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487449 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5b361c60-8056-4c29-a634-97c8a1c58f67-certs\") pod \"machine-config-server-xxbtt\" (UID: \"5b361c60-8056-4c29-a634-97c8a1c58f67\") " pod="openshift-machine-config-operator/machine-config-server-xxbtt" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487470 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-webhook-cert\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487504 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5df805e2-6930-4217-adbb-8c1344b15fbb-proxy-tls\") pod \"machine-config-controller-84d6567774-wfvkx\" (UID: \"5df805e2-6930-4217-adbb-8c1344b15fbb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487528 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctkkz\" (UniqueName: \"kubernetes.io/projected/689650fc-6a12-4737-987d-1b1205e46e41-kube-api-access-ctkkz\") pod \"catalog-operator-68c6474976-rfd6l\" (UID: \"689650fc-6a12-4737-987d-1b1205e46e41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487569 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8d81122-beac-495b-9e7f-76fad0fd5fa1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zzks6\" (UID: \"f8d81122-beac-495b-9e7f-76fad0fd5fa1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487593 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5b361c60-8056-4c29-a634-97c8a1c58f67-node-bootstrap-token\") pod \"machine-config-server-xxbtt\" (UID: \"5b361c60-8056-4c29-a634-97c8a1c58f67\") " pod="openshift-machine-config-operator/machine-config-server-xxbtt" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487622 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/689650fc-6a12-4737-987d-1b1205e46e41-profile-collector-cert\") pod \"catalog-operator-68c6474976-rfd6l\" (UID: \"689650fc-6a12-4737-987d-1b1205e46e41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487661 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-config-volume\") pod \"collect-profiles-29320320-kzpbn\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487689 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sck96\" (UniqueName: \"kubernetes.io/projected/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-kube-api-access-sck96\") pod \"collect-profiles-29320320-kzpbn\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487727 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f8d81122-beac-495b-9e7f-76fad0fd5fa1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zzks6\" (UID: \"f8d81122-beac-495b-9e7f-76fad0fd5fa1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487752 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-mountpoint-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487778 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6r2j\" (UniqueName: \"kubernetes.io/projected/00c025fa-c660-4977-a0f6-bf3d1c64ccd2-kube-api-access-t6r2j\") pod \"service-ca-9c57cc56f-5fwj8\" (UID: \"00c025fa-c660-4977-a0f6-bf3d1c64ccd2\") " pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487806 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5df805e2-6930-4217-adbb-8c1344b15fbb-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wfvkx\" (UID: \"5df805e2-6930-4217-adbb-8c1344b15fbb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487831 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4-serving-cert\") pod \"service-ca-operator-777779d784-4mdfd\" (UID: \"fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487855 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7qb9\" (UniqueName: \"kubernetes.io/projected/1c38bb35-d8db-40f6-903a-10682aa99f35-kube-api-access-v7qb9\") pod \"dns-default-6lcfr\" (UID: \"1c38bb35-d8db-40f6-903a-10682aa99f35\") " pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487884 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06d3ea99-b8e6-4e46-bd04-e370f73274ef-config\") pod \"kube-apiserver-operator-766d6c64bb-6gc9d\" (UID: \"06d3ea99-b8e6-4e46-bd04-e370f73274ef\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487909 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a7250b9a-5aef-438b-b767-430287477d1e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5h6v8\" (UID: \"a7250b9a-5aef-438b-b767-430287477d1e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487934 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a7250b9a-5aef-438b-b767-430287477d1e-srv-cert\") pod \"olm-operator-6b444d44fb-5h6v8\" (UID: \"a7250b9a-5aef-438b-b767-430287477d1e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487957 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brtxg\" (UniqueName: \"kubernetes.io/projected/167864ca-6a75-4fc0-961a-9af57dcce611-kube-api-access-brtxg\") pod \"ingress-canary-khj2k\" (UID: \"167864ca-6a75-4fc0-961a-9af57dcce611\") " pod="openshift-ingress-canary/ingress-canary-khj2k" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.487979 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-tmpfs\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.488002 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f4b41c70-05a1-4342-8aee-cba6e44b89dc-metrics-certs\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.488027 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-apiservice-cert\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.488054 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f4b41c70-05a1-4342-8aee-cba6e44b89dc-default-certificate\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.488078 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-proxy-tls\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.488103 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1c38bb35-d8db-40f6-903a-10682aa99f35-metrics-tls\") pod \"dns-default-6lcfr\" (UID: \"1c38bb35-d8db-40f6-903a-10682aa99f35\") " pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.488928 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-registration-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.489486 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-csi-data-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.489556 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-plugins-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.489580 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4b41c70-05a1-4342-8aee-cba6e44b89dc-service-ca-bundle\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.489626 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cc78ba14-6088-4423-932c-9377a1c99a2d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fkmsl\" (UID: \"cc78ba14-6088-4423-932c-9377a1c99a2d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.490009 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4-config\") pod \"service-ca-operator-777779d784-4mdfd\" (UID: \"fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.490724 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-socket-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.490919 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1c38bb35-d8db-40f6-903a-10682aa99f35-config-volume\") pod \"dns-default-6lcfr\" (UID: \"1c38bb35-d8db-40f6-903a-10682aa99f35\") " pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.495230 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf"] Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.495923 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/00c025fa-c660-4977-a0f6-bf3d1c64ccd2-signing-cabundle\") pod \"service-ca-9c57cc56f-5fwj8\" (UID: \"00c025fa-c660-4977-a0f6-bf3d1c64ccd2\") " pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.496242 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tmk5l\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.496633 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8d81122-beac-495b-9e7f-76fad0fd5fa1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zzks6\" (UID: \"f8d81122-beac-495b-9e7f-76fad0fd5fa1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.496894 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a7250b9a-5aef-438b-b767-430287477d1e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5h6v8\" (UID: \"a7250b9a-5aef-438b-b767-430287477d1e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.496963 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-mountpoint-dir\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.497650 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.497750 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/00e1ebdd-3664-401d-b7bf-a9f786d9b6d2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p7cc4\" (UID: \"00e1ebdd-3664-401d-b7bf-a9f786d9b6d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.498428 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8d81122-beac-495b-9e7f-76fad0fd5fa1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zzks6\" (UID: \"f8d81122-beac-495b-9e7f-76fad0fd5fa1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.498862 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-images\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: E0930 08:05:19.499140 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:19.999126025 +0000 UTC m=+143.451325292 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.500487 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f4b41c70-05a1-4342-8aee-cba6e44b89dc-stats-auth\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.500647 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5df805e2-6930-4217-adbb-8c1344b15fbb-proxy-tls\") pod \"machine-config-controller-84d6567774-wfvkx\" (UID: \"5df805e2-6930-4217-adbb-8c1344b15fbb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.500904 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5971693-8b44-4a2a-bd8f-7ee41e0f69cb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-blbjc\" (UID: \"c5971693-8b44-4a2a-bd8f-7ee41e0f69cb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.501411 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5df805e2-6930-4217-adbb-8c1344b15fbb-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wfvkx\" (UID: \"5df805e2-6930-4217-adbb-8c1344b15fbb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.501578 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06d3ea99-b8e6-4e46-bd04-e370f73274ef-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6gc9d\" (UID: \"06d3ea99-b8e6-4e46-bd04-e370f73274ef\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.501879 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06d3ea99-b8e6-4e46-bd04-e370f73274ef-config\") pod \"kube-apiserver-operator-766d6c64bb-6gc9d\" (UID: \"06d3ea99-b8e6-4e46-bd04-e370f73274ef\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.501934 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1c38bb35-d8db-40f6-903a-10682aa99f35-metrics-tls\") pod \"dns-default-6lcfr\" (UID: \"1c38bb35-d8db-40f6-903a-10682aa99f35\") " pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.502244 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-tmpfs\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.502806 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tmk5l\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.502886 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-config-volume\") pod \"collect-profiles-29320320-kzpbn\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.502979 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/00c025fa-c660-4977-a0f6-bf3d1c64ccd2-signing-key\") pod \"service-ca-9c57cc56f-5fwj8\" (UID: \"00c025fa-c660-4977-a0f6-bf3d1c64ccd2\") " pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.503522 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-secret-volume\") pod \"collect-profiles-29320320-kzpbn\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.503643 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5b361c60-8056-4c29-a634-97c8a1c58f67-node-bootstrap-token\") pod \"machine-config-server-xxbtt\" (UID: \"5b361c60-8056-4c29-a634-97c8a1c58f67\") " pod="openshift-machine-config-operator/machine-config-server-xxbtt" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.504256 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a7250b9a-5aef-438b-b767-430287477d1e-srv-cert\") pod \"olm-operator-6b444d44fb-5h6v8\" (UID: \"a7250b9a-5aef-438b-b767-430287477d1e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.504484 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-webhook-cert\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.504799 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5b361c60-8056-4c29-a634-97c8a1c58f67-certs\") pod \"machine-config-server-xxbtt\" (UID: \"5b361c60-8056-4c29-a634-97c8a1c58f67\") " pod="openshift-machine-config-operator/machine-config-server-xxbtt" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.504845 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4-serving-cert\") pod \"service-ca-operator-777779d784-4mdfd\" (UID: \"fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.505289 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f4b41c70-05a1-4342-8aee-cba6e44b89dc-default-certificate\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.505401 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/689650fc-6a12-4737-987d-1b1205e46e41-profile-collector-cert\") pod \"catalog-operator-68c6474976-rfd6l\" (UID: \"689650fc-6a12-4737-987d-1b1205e46e41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.505939 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f4b41c70-05a1-4342-8aee-cba6e44b89dc-metrics-certs\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.506115 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-apiservice-cert\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.506791 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-proxy-tls\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.506889 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/167864ca-6a75-4fc0-961a-9af57dcce611-cert\") pod \"ingress-canary-khj2k\" (UID: \"167864ca-6a75-4fc0-961a-9af57dcce611\") " pod="openshift-ingress-canary/ingress-canary-khj2k" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.512124 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/689650fc-6a12-4737-987d-1b1205e46e41-srv-cert\") pod \"catalog-operator-68c6474976-rfd6l\" (UID: \"689650fc-6a12-4737-987d-1b1205e46e41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.515511 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6v92\" (UniqueName: \"kubernetes.io/projected/5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a-kube-api-access-q6v92\") pod \"openshift-apiserver-operator-796bbdcf4f-s2vww\" (UID: \"5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.531423 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxxj2\" (UniqueName: \"kubernetes.io/projected/7f643a6b-bd6f-40f8-8857-ac09ac806b2f-kube-api-access-wxxj2\") pod \"authentication-operator-69f744f599-zhxt6\" (UID: \"7f643a6b-bd6f-40f8-8857-ac09ac806b2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: W0930 08:05:19.532162 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod479ebf74_62ea_4516_8c29_8d9cdf128fb4.slice/crio-2f08d9d1ab22304f057a606b0e51d8555b94d249423e33ec48c1733788c6ef26 WatchSource:0}: Error finding container 2f08d9d1ab22304f057a606b0e51d8555b94d249423e33ec48c1733788c6ef26: Status 404 returned error can't find the container with id 2f08d9d1ab22304f057a606b0e51d8555b94d249423e33ec48c1733788c6ef26 Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.536497 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5w577"] Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.556416 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d60f7e34-64f7-4364-b1a3-de505ad145df-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.572438 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj"] Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.572687 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-47kbw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.573132 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj8x6\" (UniqueName: \"kubernetes.io/projected/d60f7e34-64f7-4364-b1a3-de505ad145df-kube-api-access-jj8x6\") pod \"cluster-image-registry-operator-dc59b4c8b-cfqgb\" (UID: \"d60f7e34-64f7-4364-b1a3-de505ad145df\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.582459 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" Sep 30 08:05:19 crc kubenswrapper[4810]: W0930 08:05:19.587515 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc807324c_da31_4d36_8640_186d9cc31097.slice/crio-da89316ba053f7d20da0fa569689678c6717da4dd988d1ec16ac28b375014ef6 WatchSource:0}: Error finding container da89316ba053f7d20da0fa569689678c6717da4dd988d1ec16ac28b375014ef6: Status 404 returned error can't find the container with id da89316ba053f7d20da0fa569689678c6717da4dd988d1ec16ac28b375014ef6 Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.589045 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:19 crc kubenswrapper[4810]: E0930 08:05:19.589339 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:20.089319364 +0000 UTC m=+143.541518631 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.589738 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: E0930 08:05:19.590068 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:20.090057715 +0000 UTC m=+143.542256972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.591112 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt7c7\" (UniqueName: \"kubernetes.io/projected/6083d85f-fd9a-4df3-84c5-f6a92e84c53c-kube-api-access-rt7c7\") pod \"kube-storage-version-migrator-operator-b67b599dd-49zt7\" (UID: \"6083d85f-fd9a-4df3-84c5-f6a92e84c53c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.595702 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.612613 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs9h6\" (UniqueName: \"kubernetes.io/projected/6760d07c-3b94-492f-b58c-c61dbd1df867-kube-api-access-gs9h6\") pod \"multus-admission-controller-857f4d67dd-5lb54\" (UID: \"6760d07c-3b94-492f-b58c-c61dbd1df867\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.631433 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k965l\" (UniqueName: \"kubernetes.io/projected/626d8738-cb60-49b6-ba04-754600d28a86-kube-api-access-k965l\") pod \"machine-approver-56656f9798-rh9mj\" (UID: \"626d8738-cb60-49b6-ba04-754600d28a86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.647438 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.655778 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc4cz\" (UniqueName: \"kubernetes.io/projected/95f2618f-a0a5-43a2-b639-25944e913395-kube-api-access-zc4cz\") pod \"oauth-openshift-558db77b4-nzjhz\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.673731 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmjfj\" (UniqueName: \"kubernetes.io/projected/e57275b7-e952-4962-89fd-7950616d05f0-kube-api-access-lmjfj\") pod \"etcd-operator-b45778765-smwhg\" (UID: \"e57275b7-e952-4962-89fd-7950616d05f0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.676568 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.689708 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.695574 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.695898 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" Sep 30 08:05:19 crc kubenswrapper[4810]: E0930 08:05:19.696089 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:20.196041711 +0000 UTC m=+143.648240988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.696173 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: E0930 08:05:19.696743 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:20.196733631 +0000 UTC m=+143.648932908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.700864 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.703072 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkl2r\" (UniqueName: \"kubernetes.io/projected/f840f471-2128-47ce-9bd6-10eb24c31be5-kube-api-access-tkl2r\") pod \"openshift-controller-manager-operator-756b6f6bc6-5wczx\" (UID: \"f840f471-2128-47ce-9bd6-10eb24c31be5\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.737835 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbdjr\" (UniqueName: \"kubernetes.io/projected/bc765a10-3756-403a-8723-2cb47e961f26-kube-api-access-hbdjr\") pod \"migrator-59844c95c7-xtbw5\" (UID: \"bc765a10-3756-403a-8723-2cb47e961f26\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.765233 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp8nw\" (UniqueName: \"kubernetes.io/projected/c5971693-8b44-4a2a-bd8f-7ee41e0f69cb-kube-api-access-bp8nw\") pod \"package-server-manager-789f6589d5-blbjc\" (UID: \"c5971693-8b44-4a2a-bd8f-7ee41e0f69cb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.769672 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.779091 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vdsr\" (UniqueName: \"kubernetes.io/projected/5b361c60-8056-4c29-a634-97c8a1c58f67-kube-api-access-8vdsr\") pod \"machine-config-server-xxbtt\" (UID: \"5b361c60-8056-4c29-a634-97c8a1c58f67\") " pod="openshift-machine-config-operator/machine-config-server-xxbtt" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.786254 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.792745 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.796005 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl8vd\" (UniqueName: \"kubernetes.io/projected/bb7f5687-2b58-4f8e-9d58-51707b2e0c2c-kube-api-access-bl8vd\") pod \"machine-config-operator-74547568cd-5q5vw\" (UID: \"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.797969 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:19 crc kubenswrapper[4810]: E0930 08:05:19.798615 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:20.29859298 +0000 UTC m=+143.750792247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.810320 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.829169 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlfxs\" (UniqueName: \"kubernetes.io/projected/fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4-kube-api-access-tlfxs\") pod \"service-ca-operator-777779d784-4mdfd\" (UID: \"fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.841165 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4sbv\" (UniqueName: \"kubernetes.io/projected/5cc8817f-e25c-4a2d-aa65-2696793ff6fc-kube-api-access-z4sbv\") pod \"packageserver-d55dfcdfc-jrsk2\" (UID: \"5cc8817f-e25c-4a2d-aa65-2696793ff6fc\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.856608 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p22n\" (UniqueName: \"kubernetes.io/projected/5df805e2-6930-4217-adbb-8c1344b15fbb-kube-api-access-4p22n\") pod \"machine-config-controller-84d6567774-wfvkx\" (UID: \"5df805e2-6930-4217-adbb-8c1344b15fbb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.870303 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xxbtt" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.874217 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsqlm\" (UniqueName: \"kubernetes.io/projected/e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4-kube-api-access-tsqlm\") pod \"csi-hostpathplugin-k785t\" (UID: \"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4\") " pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.892415 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnvvq\" (UniqueName: \"kubernetes.io/projected/a7250b9a-5aef-438b-b767-430287477d1e-kube-api-access-hnvvq\") pod \"olm-operator-6b444d44fb-5h6v8\" (UID: \"a7250b9a-5aef-438b-b767-430287477d1e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.901870 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:19 crc kubenswrapper[4810]: E0930 08:05:19.902842 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:20.402755535 +0000 UTC m=+143.854954802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.905665 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.917373 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rpx5\" (UniqueName: \"kubernetes.io/projected/00e1ebdd-3664-401d-b7bf-a9f786d9b6d2-kube-api-access-5rpx5\") pod \"control-plane-machine-set-operator-78cbb6b69f-p7cc4\" (UID: \"00e1ebdd-3664-401d-b7bf-a9f786d9b6d2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.952414 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5892\" (UniqueName: \"kubernetes.io/projected/e049bda6-e38f-4357-926e-d8b767cf5ed9-kube-api-access-m5892\") pod \"marketplace-operator-79b997595-tmk5l\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.953943 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f8d81122-beac-495b-9e7f-76fad0fd5fa1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zzks6\" (UID: \"f8d81122-beac-495b-9e7f-76fad0fd5fa1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" Sep 30 08:05:19 crc kubenswrapper[4810]: W0930 08:05:19.953999 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b361c60_8056_4c29_a634_97c8a1c58f67.slice/crio-5ceaec3f087ba47862751a3c252cb39def5e52f7737475dcbf953aef93989922 WatchSource:0}: Error finding container 5ceaec3f087ba47862751a3c252cb39def5e52f7737475dcbf953aef93989922: Status 404 returned error can't find the container with id 5ceaec3f087ba47862751a3c252cb39def5e52f7737475dcbf953aef93989922 Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.962354 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" Sep 30 08:05:19 crc kubenswrapper[4810]: I0930 08:05:19.978622 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv99d\" (UniqueName: \"kubernetes.io/projected/f4b41c70-05a1-4342-8aee-cba6e44b89dc-kube-api-access-gv99d\") pod \"router-default-5444994796-555gv\" (UID: \"f4b41c70-05a1-4342-8aee-cba6e44b89dc\") " pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.004170 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:20 crc kubenswrapper[4810]: E0930 08:05:20.004793 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:20.504762519 +0000 UTC m=+143.956961786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.005864 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6r2j\" (UniqueName: \"kubernetes.io/projected/00c025fa-c660-4977-a0f6-bf3d1c64ccd2-kube-api-access-t6r2j\") pod \"service-ca-9c57cc56f-5fwj8\" (UID: \"00c025fa-c660-4977-a0f6-bf3d1c64ccd2\") " pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.008371 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.022441 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.029544 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.031081 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctkkz\" (UniqueName: \"kubernetes.io/projected/689650fc-6a12-4737-987d-1b1205e46e41-kube-api-access-ctkkz\") pod \"catalog-operator-68c6474976-rfd6l\" (UID: \"689650fc-6a12-4737-987d-1b1205e46e41\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.036095 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.043789 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.046899 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06d3ea99-b8e6-4e46-bd04-e370f73274ef-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6gc9d\" (UID: \"06d3ea99-b8e6-4e46-bd04-e370f73274ef\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.050690 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.057249 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.066469 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7qb9\" (UniqueName: \"kubernetes.io/projected/1c38bb35-d8db-40f6-903a-10682aa99f35-kube-api-access-v7qb9\") pod \"dns-default-6lcfr\" (UID: \"1c38bb35-d8db-40f6-903a-10682aa99f35\") " pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.074117 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.079769 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.084429 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-47kbw"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.094102 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.108920 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.108922 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brtxg\" (UniqueName: \"kubernetes.io/projected/167864ca-6a75-4fc0-961a-9af57dcce611-kube-api-access-brtxg\") pod \"ingress-canary-khj2k\" (UID: \"167864ca-6a75-4fc0-961a-9af57dcce611\") " pod="openshift-ingress-canary/ingress-canary-khj2k" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.109469 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.120296 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sck96\" (UniqueName: \"kubernetes.io/projected/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-kube-api-access-sck96\") pod \"collect-profiles-29320320-kzpbn\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.123175 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.124227 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-khj2k" Sep 30 08:05:20 crc kubenswrapper[4810]: E0930 08:05:20.131975 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:20.631951814 +0000 UTC m=+144.084151081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.156055 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.156288 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-k785t" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.160632 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.210631 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:20 crc kubenswrapper[4810]: E0930 08:05:20.210980 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:20.710963848 +0000 UTC m=+144.163163115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:20 crc kubenswrapper[4810]: W0930 08:05:20.255122 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode90bfa19_6003_4593_b318_31ab8ec8c44f.slice/crio-8732f4dfe1648b97324d39c96d244818dc9058a0812c5b5106e86ee8db2462b0 WatchSource:0}: Error finding container 8732f4dfe1648b97324d39c96d244818dc9058a0812c5b5106e86ee8db2462b0: Status 404 returned error can't find the container with id 8732f4dfe1648b97324d39c96d244818dc9058a0812c5b5106e86ee8db2462b0 Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.284510 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" event={"ID":"479ebf74-62ea-4516-8c29-8d9cdf128fb4","Type":"ContainerStarted","Data":"2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.284549 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" event={"ID":"479ebf74-62ea-4516-8c29-8d9cdf128fb4","Type":"ContainerStarted","Data":"2f08d9d1ab22304f057a606b0e51d8555b94d249423e33ec48c1733788c6ef26"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.285457 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.291326 4810 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-xggpf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.291367 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" podUID="479ebf74-62ea-4516-8c29-8d9cdf128fb4" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.294587 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.300313 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-49gdq" event={"ID":"31e65ba1-328b-4d4b-905c-6074ab5d1427","Type":"ContainerStarted","Data":"558dc45a511553b15d7d1cd9668341a1262a9ec3bd94a84c83e4295080e959c5"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.300360 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.300372 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-49gdq" event={"ID":"31e65ba1-328b-4d4b-905c-6074ab5d1427","Type":"ContainerStarted","Data":"3670228e815d69d16e0ce9121d3b8e4358ac9d15a40e6eb551a9c4e1f25f466f"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.304749 4810 patch_prober.go:28] interesting pod/console-operator-58897d9998-49gdq container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/readyz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.304793 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-49gdq" podUID="31e65ba1-328b-4d4b-905c-6074ab5d1427" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/readyz\": dial tcp 10.217.0.28:8443: connect: connection refused" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.311515 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:20 crc kubenswrapper[4810]: E0930 08:05:20.312179 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:20.812165719 +0000 UTC m=+144.264364986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.314942 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.333836 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" event={"ID":"c807324c-da31-4d36-8640-186d9cc31097","Type":"ContainerStarted","Data":"1c76aa841786262c7d0f6f8a917419c2faa1617a241e5c8dd842b286f0ffea2b"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.333890 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" event={"ID":"c807324c-da31-4d36-8640-186d9cc31097","Type":"ContainerStarted","Data":"e70606769de2f7e8f73c55021c06d59d6bcb168f6d328880a79ec7953157767e"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.333905 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" event={"ID":"c807324c-da31-4d36-8640-186d9cc31097","Type":"ContainerStarted","Data":"da89316ba053f7d20da0fa569689678c6717da4dd988d1ec16ac28b375014ef6"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.360034 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" event={"ID":"52e8e292-0788-4957-ac33-083fadb0f61b","Type":"ContainerStarted","Data":"013ec22f239dcbc0643de72b295404e98ccae68ab935afaadc8a284f7356b7c3"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.360107 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" event={"ID":"52e8e292-0788-4957-ac33-083fadb0f61b","Type":"ContainerStarted","Data":"27249e0371df3fe93103c2ce64415e56b880f49594e44f9c47dfc278007fa88c"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.371010 4810 generic.go:334] "Generic (PLEG): container finished" podID="cd62560a-d47e-49f2-b3cd-06568dc86f43" containerID="a841385cd609e77ba14bddef531633f04185b7625196981776936d5f9912f2a4" exitCode=0 Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.371791 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.371825 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" event={"ID":"cd62560a-d47e-49f2-b3cd-06568dc86f43","Type":"ContainerDied","Data":"a841385cd609e77ba14bddef531633f04185b7625196981776936d5f9912f2a4"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.372063 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zhxt6"] Sep 30 08:05:20 crc kubenswrapper[4810]: W0930 08:05:20.384693 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fbb38bc_cab5_483d_80f1_7d8ad9a3d14a.slice/crio-5167012582ae24e77ed6d6e718f4ca32c5e213b4a28638895c9ce1528a250332 WatchSource:0}: Error finding container 5167012582ae24e77ed6d6e718f4ca32c5e213b4a28638895c9ce1528a250332: Status 404 returned error can't find the container with id 5167012582ae24e77ed6d6e718f4ca32c5e213b4a28638895c9ce1528a250332 Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.384785 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" event={"ID":"f7201c6b-de6b-4c4e-b20b-90c59d0db7ac","Type":"ContainerStarted","Data":"10a2a4a139ed8c4aaa0129576fee64e93c9b532b26fc9b6b3051b1d6c43620f2"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.399809 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.400220 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" event={"ID":"626d8738-cb60-49b6-ba04-754600d28a86","Type":"ContainerStarted","Data":"dbbbef081edee2b1dd909d89e3457b4028ff7dc67679078fe66103f01728557d"} Sep 30 08:05:20 crc kubenswrapper[4810]: W0930 08:05:20.400763 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd60f7e34_64f7_4364_b1a3_de505ad145df.slice/crio-e3b6b52e06bcace909405e862fe03767ae5df5958fca227d2eeeda9488acd080 WatchSource:0}: Error finding container e3b6b52e06bcace909405e862fe03767ae5df5958fca227d2eeeda9488acd080: Status 404 returned error can't find the container with id e3b6b52e06bcace909405e862fe03767ae5df5958fca227d2eeeda9488acd080 Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.403863 4810 generic.go:334] "Generic (PLEG): container finished" podID="60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c" containerID="207f2129fdc1334208b43126f8db14743050a9c3939c57aec67b1ef5efbd12d2" exitCode=0 Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.403966 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" event={"ID":"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c","Type":"ContainerDied","Data":"207f2129fdc1334208b43126f8db14743050a9c3939c57aec67b1ef5efbd12d2"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.405229 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xxbtt" event={"ID":"5b361c60-8056-4c29-a634-97c8a1c58f67","Type":"ContainerStarted","Data":"5ceaec3f087ba47862751a3c252cb39def5e52f7737475dcbf953aef93989922"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.405966 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-555gv" event={"ID":"f4b41c70-05a1-4342-8aee-cba6e44b89dc","Type":"ContainerStarted","Data":"53caf7a99081cb842762ac68f9dfbb481cef84eb7db12e890c7f629256904be3"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.407235 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5w577" event={"ID":"64fd38ac-d0ef-430c-aed5-5a4e4944986c","Type":"ContainerStarted","Data":"9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.407316 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5w577" event={"ID":"64fd38ac-d0ef-430c-aed5-5a4e4944986c","Type":"ContainerStarted","Data":"030ee37eb2e6e0b30ca1099988ce7322f75b321380bcb63716be10c79f73e47d"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.411493 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" event={"ID":"3a1774b8-c928-4843-85ba-e02066fd0f5c","Type":"ContainerStarted","Data":"7b8942d6eba7399d9adddd47acb852fb29a8e68571a690af881ccfbdfd333f1d"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.411546 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" event={"ID":"3a1774b8-c928-4843-85ba-e02066fd0f5c","Type":"ContainerStarted","Data":"521b42c220e3c89660279b7edbe985df1c72e87c9d5fc00532dd4f3c9dcb03ed"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.411557 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" event={"ID":"3a1774b8-c928-4843-85ba-e02066fd0f5c","Type":"ContainerStarted","Data":"a3d86e42026332798b188251a38e2c040657c816522780849121364de6e2721d"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.412217 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:20 crc kubenswrapper[4810]: E0930 08:05:20.413852 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:20.913833403 +0000 UTC m=+144.366032670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.425109 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" event={"ID":"80a2e9d7-a1ec-436d-8634-89c763644dec","Type":"ContainerStarted","Data":"638c48838c29dccb16889232e47ee51085f0be306b2c6857be7f7195f2fa6135"} Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.426233 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.429709 4810 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-qdcml container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.429753 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" podUID="80a2e9d7-a1ec-436d-8634-89c763644dec" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.512904 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.513921 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:20 crc kubenswrapper[4810]: E0930 08:05:20.518244 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.018230644 +0000 UTC m=+144.470429911 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.575358 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5lb54"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.575481 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-smwhg"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.601207 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.608504 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nzjhz"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.608553 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc"] Sep 30 08:05:20 crc kubenswrapper[4810]: W0930 08:05:20.609790 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode57275b7_e952_4962_89fd_7950616d05f0.slice/crio-e638cb1181b75a44fe36e9c0f6b066c7f459ab705ea68500fae2a4dce5f73885 WatchSource:0}: Error finding container e638cb1181b75a44fe36e9c0f6b066c7f459ab705ea68500fae2a4dce5f73885: Status 404 returned error can't find the container with id e638cb1181b75a44fe36e9c0f6b066c7f459ab705ea68500fae2a4dce5f73885 Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.615066 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:20 crc kubenswrapper[4810]: E0930 08:05:20.615514 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.115498434 +0000 UTC m=+144.567697701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:20 crc kubenswrapper[4810]: W0930 08:05:20.652966 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5971693_8b44_4a2a_bd8f_7ee41e0f69cb.slice/crio-7b952f067ed8c00687c856c858ed4ab268ccd3239fb7a63df991a6b35c0d5f65 WatchSource:0}: Error finding container 7b952f067ed8c00687c856c858ed4ab268ccd3239fb7a63df991a6b35c0d5f65: Status 404 returned error can't find the container with id 7b952f067ed8c00687c856c858ed4ab268ccd3239fb7a63df991a6b35c0d5f65 Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.711763 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.722697 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:20 crc kubenswrapper[4810]: E0930 08:05:20.723011 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.222999973 +0000 UTC m=+144.675199230 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.751854 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.751902 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.775362 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l"] Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.824888 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:20 crc kubenswrapper[4810]: E0930 08:05:20.825198 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.325183351 +0000 UTC m=+144.777382618 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.926071 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:20 crc kubenswrapper[4810]: E0930 08:05:20.926918 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.426902707 +0000 UTC m=+144.879101974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.998220 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-5w577" podStartSLOduration=123.998201473 podStartE2EDuration="2m3.998201473s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:20.997455651 +0000 UTC m=+144.449654918" watchObservedRunningTime="2025-09-30 08:05:20.998201473 +0000 UTC m=+144.450400730" Sep 30 08:05:20 crc kubenswrapper[4810]: I0930 08:05:20.999216 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd"] Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.003588 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx"] Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.027845 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.028064 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.528042016 +0000 UTC m=+144.980241283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.028309 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.028632 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.528625513 +0000 UTC m=+144.980824780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: W0930 08:05:21.050232 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf840f471_2128_47ce_9bd6_10eb24c31be5.slice/crio-67bcab5e7202f18cb9da412f01160d18dca56b15825887a790e751f99cf1cc2c WatchSource:0}: Error finding container 67bcab5e7202f18cb9da412f01160d18dca56b15825887a790e751f99cf1cc2c: Status 404 returned error can't find the container with id 67bcab5e7202f18cb9da412f01160d18dca56b15825887a790e751f99cf1cc2c Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.134226 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.134378 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.634362372 +0000 UTC m=+145.086561639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.135311 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.135642 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.635634998 +0000 UTC m=+145.087834265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.198908 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-nj7lz" podStartSLOduration=123.198889116 podStartE2EDuration="2m3.198889116s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:21.1586927 +0000 UTC m=+144.610891957" watchObservedRunningTime="2025-09-30 08:05:21.198889116 +0000 UTC m=+144.651088383" Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.289077 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.293388 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.793358036 +0000 UTC m=+145.245557303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.305879 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8"] Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.320816 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.321400 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.821372798 +0000 UTC m=+145.273572065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.339166 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" podStartSLOduration=124.33913539 podStartE2EDuration="2m4.33913539s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:21.332639757 +0000 UTC m=+144.784839024" watchObservedRunningTime="2025-09-30 08:05:21.33913539 +0000 UTC m=+144.791334657" Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.396290 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5fwj8"] Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.417120 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" podStartSLOduration=123.417100925 podStartE2EDuration="2m3.417100925s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:21.40952161 +0000 UTC m=+144.861720877" watchObservedRunningTime="2025-09-30 08:05:21.417100925 +0000 UTC m=+144.869300192" Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.430893 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.431927 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:21.931898473 +0000 UTC m=+145.384097740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.540895 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.541327 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:22.041312616 +0000 UTC m=+145.493511873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.584588 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" event={"ID":"5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a","Type":"ContainerStarted","Data":"5167012582ae24e77ed6d6e718f4ca32c5e213b4a28638895c9ce1528a250332"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.589938 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" event={"ID":"f840f471-2128-47ce-9bd6-10eb24c31be5","Type":"ContainerStarted","Data":"67bcab5e7202f18cb9da412f01160d18dca56b15825887a790e751f99cf1cc2c"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.615048 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2"] Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.616025 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-47kbw" event={"ID":"e90bfa19-6003-4593-b318-31ab8ec8c44f","Type":"ContainerStarted","Data":"9bb441593cd8a2b5b44bc35dc245855b54c73598955be9a6028f81567ae3733e"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.616075 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-47kbw" event={"ID":"e90bfa19-6003-4593-b318-31ab8ec8c44f","Type":"ContainerStarted","Data":"8732f4dfe1648b97324d39c96d244818dc9058a0812c5b5106e86ee8db2462b0"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.617610 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-47kbw" Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.619566 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6"] Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.627203 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" event={"ID":"cc78ba14-6088-4423-932c-9377a1c99a2d","Type":"ContainerStarted","Data":"989862fdd84e8ae698eafddaf184b63bfe5b1097c9626856a97d6716852a42ed"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.640714 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-47kbw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.640770 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-47kbw" podUID="e90bfa19-6003-4593-b318-31ab8ec8c44f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.641668 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.642489 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:22.142460075 +0000 UTC m=+145.594659432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.712308 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" event={"ID":"5df805e2-6930-4217-adbb-8c1344b15fbb","Type":"ContainerStarted","Data":"a4d83d98b2185abdbac3eb1c86c328353a200622df6d9d4312715904f5482f09"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.722587 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-49gdq" podStartSLOduration=124.72256957 podStartE2EDuration="2m4.72256957s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:21.720381728 +0000 UTC m=+145.172580995" watchObservedRunningTime="2025-09-30 08:05:21.72256957 +0000 UTC m=+145.174768837" Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.724319 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" event={"ID":"95f2618f-a0a5-43a2-b639-25944e913395","Type":"ContainerStarted","Data":"996c3b09d20d02362147173fff81f2b0299802853d5835a16b719dcf96889e34"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.743038 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.743359 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:22.243346667 +0000 UTC m=+145.695545934 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: W0930 08:05:21.744975 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8d81122_beac_495b_9e7f_76fad0fd5fa1.slice/crio-98a106fc6eec02d9f78837575f49cf583adb5a42bece0a60f1e0783a58bfa22a WatchSource:0}: Error finding container 98a106fc6eec02d9f78837575f49cf583adb5a42bece0a60f1e0783a58bfa22a: Status 404 returned error can't find the container with id 98a106fc6eec02d9f78837575f49cf583adb5a42bece0a60f1e0783a58bfa22a Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.745008 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" event={"ID":"52e8e292-0788-4957-ac33-083fadb0f61b","Type":"ContainerStarted","Data":"59d4e81da444b35036062f6d5c7c7585b47ed7c173d27fdb15929534471944eb"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.774948 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-khj2k"] Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.776868 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" event={"ID":"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c","Type":"ContainerStarted","Data":"b994bf4709313dbcaee37fd989f6b61f86c35c18fc282fe4b62375b884d44ad9"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.778995 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" event={"ID":"d60f7e34-64f7-4364-b1a3-de505ad145df","Type":"ContainerStarted","Data":"e3b6b52e06bcace909405e862fe03767ae5df5958fca227d2eeeda9488acd080"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.790771 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mnhwj" podStartSLOduration=124.790753277 podStartE2EDuration="2m4.790753277s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:21.789669387 +0000 UTC m=+145.241868654" watchObservedRunningTime="2025-09-30 08:05:21.790753277 +0000 UTC m=+145.242952544" Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.810243 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" event={"ID":"fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4","Type":"ContainerStarted","Data":"127f31317fe1f77a390dad4c41700736136b915030a6868096d741f8da5600ef"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.811459 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tmk5l"] Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.830111 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" event={"ID":"689650fc-6a12-4737-987d-1b1205e46e41","Type":"ContainerStarted","Data":"20264f41ff647450007a9733f8c3f0ccb642c114993d5060400f13f3dc1582a5"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.839078 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6lcfr"] Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.840696 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-k785t"] Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.847211 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.848342 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:22.348243003 +0000 UTC m=+145.800442270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.849931 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn"] Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.868413 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" event={"ID":"6083d85f-fd9a-4df3-84c5-f6a92e84c53c","Type":"ContainerStarted","Data":"68b9e1218208c164720aae6235efe286d0036d08fbd41f75a101c175fe137383"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.876214 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4" event={"ID":"00e1ebdd-3664-401d-b7bf-a9f786d9b6d2","Type":"ContainerStarted","Data":"62c23a7519fd45cf8a5b4dd4f870b95e4cbe2a1d2b6be310507723cf8aab5151"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.902890 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" event={"ID":"e57275b7-e952-4962-89fd-7950616d05f0","Type":"ContainerStarted","Data":"e638cb1181b75a44fe36e9c0f6b066c7f459ab705ea68500fae2a4dce5f73885"} Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.928292 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" event={"ID":"626d8738-cb60-49b6-ba04-754600d28a86","Type":"ContainerStarted","Data":"b955d7104943a86cb9d5c2199d2cc60f0d2df5ee2a52728d08c8f05d78424ec6"} Sep 30 08:05:21 crc kubenswrapper[4810]: W0930 08:05:21.944695 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7bf6704_7c12_4f9d_b0bf_e1592fae2ec4.slice/crio-e663a674fa6d6672a0732b3bca46bf12c07c72f39a01f1f0154a3c58fe0d3beb WatchSource:0}: Error finding container e663a674fa6d6672a0732b3bca46bf12c07c72f39a01f1f0154a3c58fe0d3beb: Status 404 returned error can't find the container with id e663a674fa6d6672a0732b3bca46bf12c07c72f39a01f1f0154a3c58fe0d3beb Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.968541 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:21 crc kubenswrapper[4810]: E0930 08:05:21.968930 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:22.468917403 +0000 UTC m=+145.921116670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:21 crc kubenswrapper[4810]: I0930 08:05:21.969938 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d"] Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.006357 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5" event={"ID":"bc765a10-3756-403a-8723-2cb47e961f26","Type":"ContainerStarted","Data":"55405ce2778c3742aa583b3ae94c0716a5e22bfab454f9e36a36a1f11a4cc92f"} Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.013933 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcsj5" podStartSLOduration=125.013911425 podStartE2EDuration="2m5.013911425s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:21.995090203 +0000 UTC m=+145.447289470" watchObservedRunningTime="2025-09-30 08:05:22.013911425 +0000 UTC m=+145.466110692" Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.050995 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" event={"ID":"6760d07c-3b94-492f-b58c-c61dbd1df867","Type":"ContainerStarted","Data":"5a85daf83934379e429c8477ffddc0cd7a5ba9c6b0485e2f0f70210c9aeeb6ab"} Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.070798 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:22 crc kubenswrapper[4810]: E0930 08:05:22.071156 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:22.571139313 +0000 UTC m=+146.023338580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.074324 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xxbtt" event={"ID":"5b361c60-8056-4c29-a634-97c8a1c58f67","Type":"ContainerStarted","Data":"da8acb33073f8cf725512cccb3ca3432bdcfe12e143ba60aba1d04e65021f6ae"} Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.117105 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" event={"ID":"f4a58c1d-0d71-4853-b894-e14efec97e65","Type":"ContainerStarted","Data":"9ba9996afb713de565f46dd615dc6059f7953e6f9fedcd7a54804fca2ebf7629"} Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.117152 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" event={"ID":"f4a58c1d-0d71-4853-b894-e14efec97e65","Type":"ContainerStarted","Data":"141a421f2e7871b86d2133be1444965a5914f98e9c52afcd931ce804358eac64"} Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.182085 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:22 crc kubenswrapper[4810]: E0930 08:05:22.182587 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:22.682567513 +0000 UTC m=+146.134766780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.183726 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" event={"ID":"7f643a6b-bd6f-40f8-8857-ac09ac806b2f","Type":"ContainerStarted","Data":"7607db86d0f81e120016fa5da4b8e93d56d9c4dcc8b70e2a82e951141f5439fc"} Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.183759 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" event={"ID":"c5971693-8b44-4a2a-bd8f-7ee41e0f69cb","Type":"ContainerStarted","Data":"7b952f067ed8c00687c856c858ed4ab268ccd3239fb7a63df991a6b35c0d5f65"} Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.203862 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-555gv" event={"ID":"f4b41c70-05a1-4342-8aee-cba6e44b89dc","Type":"ContainerStarted","Data":"57c00348aab388ba146758021b88881dd93361883248e45f26df867ea3c66215"} Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.218951 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.224845 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.232248 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-49gdq" Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.269805 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-555gv" podStartSLOduration=125.269778258 podStartE2EDuration="2m5.269778258s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:22.250429821 +0000 UTC m=+145.702629088" watchObservedRunningTime="2025-09-30 08:05:22.269778258 +0000 UTC m=+145.721977535" Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.271801 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-p8cxw" podStartSLOduration=125.271790095 podStartE2EDuration="2m5.271790095s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:22.268404689 +0000 UTC m=+145.720603956" watchObservedRunningTime="2025-09-30 08:05:22.271790095 +0000 UTC m=+145.723989362" Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.283853 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:22 crc kubenswrapper[4810]: E0930 08:05:22.284012 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:22.78397899 +0000 UTC m=+146.236178267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.284394 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:22 crc kubenswrapper[4810]: E0930 08:05:22.286806 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:22.786792289 +0000 UTC m=+146.238991556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.314365 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-xxbtt" podStartSLOduration=5.314341698 podStartE2EDuration="5.314341698s" podCreationTimestamp="2025-09-30 08:05:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:22.301705541 +0000 UTC m=+145.753904828" watchObservedRunningTime="2025-09-30 08:05:22.314341698 +0000 UTC m=+145.766540955" Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.343378 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-47kbw" podStartSLOduration=125.343350718 podStartE2EDuration="2m5.343350718s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:22.338543152 +0000 UTC m=+145.790742419" watchObservedRunningTime="2025-09-30 08:05:22.343350718 +0000 UTC m=+145.795549985" Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.388774 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:22 crc kubenswrapper[4810]: E0930 08:05:22.389140 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:22.889120962 +0000 UTC m=+146.341320229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.412825 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" podStartSLOduration=125.412799881 podStartE2EDuration="2m5.412799881s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:22.411643759 +0000 UTC m=+145.863843026" watchObservedRunningTime="2025-09-30 08:05:22.412799881 +0000 UTC m=+145.864999148" Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.491012 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:22 crc kubenswrapper[4810]: E0930 08:05:22.491381 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:22.991368092 +0000 UTC m=+146.443567359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.592108 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:22 crc kubenswrapper[4810]: E0930 08:05:22.592575 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.092543162 +0000 UTC m=+146.544742429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.698402 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:22 crc kubenswrapper[4810]: E0930 08:05:22.698856 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.198838597 +0000 UTC m=+146.651037864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.801426 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:22 crc kubenswrapper[4810]: E0930 08:05:22.801556 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.301534481 +0000 UTC m=+146.753733748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.801810 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:22 crc kubenswrapper[4810]: E0930 08:05:22.802252 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.302243701 +0000 UTC m=+146.754442978 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:22 crc kubenswrapper[4810]: I0930 08:05:22.905088 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:22 crc kubenswrapper[4810]: E0930 08:05:22.905499 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.405483369 +0000 UTC m=+146.857682636 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.006608 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.007003 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.506986219 +0000 UTC m=+146.959185486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.010240 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.021477 4810 patch_prober.go:28] interesting pod/router-default-5444994796-555gv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 08:05:23 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Sep 30 08:05:23 crc kubenswrapper[4810]: [+]process-running ok Sep 30 08:05:23 crc kubenswrapper[4810]: healthz check failed Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.021743 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-555gv" podUID="f4b41c70-05a1-4342-8aee-cba6e44b89dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.109639 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.119112 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.619083068 +0000 UTC m=+147.071282335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.120558 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.120967 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.62095415 +0000 UTC m=+147.073153417 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.221948 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.222150 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.72212497 +0000 UTC m=+147.174324227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.222556 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" event={"ID":"f8d81122-beac-495b-9e7f-76fad0fd5fa1","Type":"ContainerStarted","Data":"5c56348bee8c6c7ba7ad9cd526c8d7dc688b0d800a195850805fcbbe103049fd"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.222603 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" event={"ID":"f8d81122-beac-495b-9e7f-76fad0fd5fa1","Type":"ContainerStarted","Data":"98a106fc6eec02d9f78837575f49cf583adb5a42bece0a60f1e0783a58bfa22a"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.224712 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.225974 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.725961819 +0000 UTC m=+147.178161086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.238210 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" event={"ID":"d60f7e34-64f7-4364-b1a3-de505ad145df","Type":"ContainerStarted","Data":"64540cd3e367044c92e6f90658c1112be84180cd218f35b8d0734adfc06f5ac6"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.241437 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" event={"ID":"95f2618f-a0a5-43a2-b639-25944e913395","Type":"ContainerStarted","Data":"aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.242333 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.245868 4810 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nzjhz container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" start-of-body= Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.246018 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" podUID="95f2618f-a0a5-43a2-b639-25944e913395" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.247882 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" event={"ID":"00c025fa-c660-4977-a0f6-bf3d1c64ccd2","Type":"ContainerStarted","Data":"7741b2394aa94e8e9db9563c1011c265c3f1d48875171598b6b9309e375d5d96"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.247940 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" event={"ID":"00c025fa-c660-4977-a0f6-bf3d1c64ccd2","Type":"ContainerStarted","Data":"e50819fa91c150bdbfb9ec9e19a133c2fcd3f9201c26f5f02e38d1d0d1f32a98"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.261924 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-khj2k" event={"ID":"167864ca-6a75-4fc0-961a-9af57dcce611","Type":"ContainerStarted","Data":"4f35c233b696021506777d88a415b4a98bab625bf8ee803b111f92147bf8bdcb"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.261972 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-khj2k" event={"ID":"167864ca-6a75-4fc0-961a-9af57dcce611","Type":"ContainerStarted","Data":"1d9af83dd89e90cc058c58eff3aaab00de16957bfbbff126c8033fc9085f4a1b"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.277122 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" event={"ID":"a7250b9a-5aef-438b-b767-430287477d1e","Type":"ContainerStarted","Data":"314e87872dcd3f02a16131d9378d199a3562af05031c8b86c5ceab293d316b85"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.277172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" event={"ID":"a7250b9a-5aef-438b-b767-430287477d1e","Type":"ContainerStarted","Data":"5bf6c1b3e827d34dbab0b5b3fba8292e06e5f28d022a77ab94460f1327223e49"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.278451 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.280221 4810 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-5h6v8 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.280280 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" podUID="a7250b9a-5aef-438b-b767-430287477d1e" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.295285 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" podStartSLOduration=126.295248938 podStartE2EDuration="2m6.295248938s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.294890847 +0000 UTC m=+146.747090114" watchObservedRunningTime="2025-09-30 08:05:23.295248938 +0000 UTC m=+146.747448205" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.296846 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zzks6" podStartSLOduration=125.296839973 podStartE2EDuration="2m5.296839973s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.246692475 +0000 UTC m=+146.698891742" watchObservedRunningTime="2025-09-30 08:05:23.296839973 +0000 UTC m=+146.749039240" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.322519 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" event={"ID":"fc254f56-d9a3-49e6-92f4-8bc9ef7ecef4","Type":"ContainerStarted","Data":"8532682429e7a0f140b087a773c549222d421da89e5d98e08ce07f0c5d56a8f8"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.326177 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.328552 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" event={"ID":"e57275b7-e952-4962-89fd-7950616d05f0","Type":"ContainerStarted","Data":"434c2146d31c3835b98bb1c4c8744798f22d7eb6cec2035fdf815996bd5bd970"} Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.328903 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.828884188 +0000 UTC m=+147.281083455 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.332395 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-cfqgb" podStartSLOduration=126.332383827 podStartE2EDuration="2m6.332383827s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.330258617 +0000 UTC m=+146.782457884" watchObservedRunningTime="2025-09-30 08:05:23.332383827 +0000 UTC m=+146.784583094" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.343782 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" event={"ID":"6760d07c-3b94-492f-b58c-c61dbd1df867","Type":"ContainerStarted","Data":"2a4281b4daad6186ddd9ed96165ce8c8f5106914b5f29c999dfe1e1d81092bfd"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.359117 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zhxt6" event={"ID":"7f643a6b-bd6f-40f8-8857-ac09ac806b2f","Type":"ContainerStarted","Data":"8f3d362e5f0a5eb60bb077a7cddd4f8263917181358af9b2c452a347f1ff9d27"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.361218 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" podStartSLOduration=125.361205052 podStartE2EDuration="2m5.361205052s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.360389239 +0000 UTC m=+146.812588506" watchObservedRunningTime="2025-09-30 08:05:23.361205052 +0000 UTC m=+146.813404309" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.373122 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" event={"ID":"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6","Type":"ContainerStarted","Data":"26b716174b8617ef9307046c3f0d04ab2fdc57b54a9b7691dae85ca34cd14319"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.373194 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" event={"ID":"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6","Type":"ContainerStarted","Data":"cc7789b59dc790bee1bc7496907de5f4e19da5871b997c4f48aa89d0408b797e"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.392332 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" event={"ID":"cc78ba14-6088-4423-932c-9377a1c99a2d","Type":"ContainerStarted","Data":"a64c819646f9bef0bdaa15bdaf400a034a229fce2b798643087cebc8933d7353"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.408154 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" event={"ID":"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c","Type":"ContainerStarted","Data":"f51e5d118a2130a4d21a6c29da7b297fa51c68f85aeff2bc2cece37c14078973"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.408222 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" event={"ID":"bb7f5687-2b58-4f8e-9d58-51707b2e0c2c","Type":"ContainerStarted","Data":"93c9997c4a8009e00729730767e7ad6b32dbfa8e98e9c92b810c1d23b536cba4"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.449029 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.462068 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:23.962041163 +0000 UTC m=+147.414240430 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.466910 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4" event={"ID":"00e1ebdd-3664-401d-b7bf-a9f786d9b6d2","Type":"ContainerStarted","Data":"2de72f57ee22b37db44a493839f2477f7f16529125f69757189dc76603613f1e"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.478615 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-khj2k" podStartSLOduration=7.478593001 podStartE2EDuration="7.478593001s" podCreationTimestamp="2025-09-30 08:05:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.472576121 +0000 UTC m=+146.924775388" watchObservedRunningTime="2025-09-30 08:05:23.478593001 +0000 UTC m=+146.930792268" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.481848 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4mdfd" podStartSLOduration=125.48141203 podStartE2EDuration="2m5.48141203s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.397725804 +0000 UTC m=+146.849925071" watchObservedRunningTime="2025-09-30 08:05:23.48141203 +0000 UTC m=+146.933611287" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.491631 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-5fwj8" podStartSLOduration=125.491606658 podStartE2EDuration="2m5.491606658s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.490223789 +0000 UTC m=+146.942423056" watchObservedRunningTime="2025-09-30 08:05:23.491606658 +0000 UTC m=+146.943805925" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.500191 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" event={"ID":"06d3ea99-b8e6-4e46-bd04-e370f73274ef","Type":"ContainerStarted","Data":"31cb25d926a1bedb90412888f5333c6cf82b3ae541a1518f578a02722a066dad"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.529701 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" event={"ID":"5df805e2-6930-4217-adbb-8c1344b15fbb","Type":"ContainerStarted","Data":"1858c74a7afd57ae59ca528118307d632452b1266560b8c5c976e50398ddfd8b"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.529919 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" event={"ID":"5df805e2-6930-4217-adbb-8c1344b15fbb","Type":"ContainerStarted","Data":"2972da716314d4719f095880ed3792b98964fbb2735d6d4241718d67e3ccda31"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.545578 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5" event={"ID":"bc765a10-3756-403a-8723-2cb47e961f26","Type":"ContainerStarted","Data":"b838576943ff80bb9e078415c4e66099c279a41128edc654b23522563a05cfd1"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.545624 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5" event={"ID":"bc765a10-3756-403a-8723-2cb47e961f26","Type":"ContainerStarted","Data":"993731a773f26f484b3d97f5972e08fcb8e56ea399831797919865634990ab24"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.554243 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.555549 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.055529696 +0000 UTC m=+147.507728963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.569966 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wfvkx" podStartSLOduration=125.569945813 podStartE2EDuration="2m5.569945813s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.56946983 +0000 UTC m=+147.021669097" watchObservedRunningTime="2025-09-30 08:05:23.569945813 +0000 UTC m=+147.022145080" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.570531 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" event={"ID":"689650fc-6a12-4737-987d-1b1205e46e41","Type":"ContainerStarted","Data":"46551cefce9e4239b5ca9c332660b3d1b77ef3a3a0dcad840767674fbad874c3"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.570872 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-smwhg" podStartSLOduration=126.57086300899999 podStartE2EDuration="2m6.570863009s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.533827392 +0000 UTC m=+146.986026659" watchObservedRunningTime="2025-09-30 08:05:23.570863009 +0000 UTC m=+147.023062276" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.572445 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.581813 4810 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-rfd6l container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.582043 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" podUID="689650fc-6a12-4737-987d-1b1205e46e41" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.582416 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" event={"ID":"e049bda6-e38f-4357-926e-d8b767cf5ed9","Type":"ContainerStarted","Data":"f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.582466 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" event={"ID":"e049bda6-e38f-4357-926e-d8b767cf5ed9","Type":"ContainerStarted","Data":"620392ec040254ddb29445b086753d82ccea05fa80f3226fa5c798e8531dfe29"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.589173 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.594027 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" event={"ID":"5fbb38bc-cab5-483d-80f1-7d8ad9a3d14a","Type":"ContainerStarted","Data":"a2954f654d3597922bc70bc347e164557cf9d85a13fd0ff6395739dc4cf8fb2e"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.591740 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fkmsl" podStartSLOduration=125.591721659 podStartE2EDuration="2m5.591721659s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.591144352 +0000 UTC m=+147.043343619" watchObservedRunningTime="2025-09-30 08:05:23.591721659 +0000 UTC m=+147.043920926" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.591840 4810 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tmk5l container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.595129 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" podUID="e049bda6-e38f-4357-926e-d8b767cf5ed9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.596417 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" event={"ID":"f840f471-2128-47ce-9bd6-10eb24c31be5","Type":"ContainerStarted","Data":"dd59f2e0b4bfd9967eec146726d9e6df6162598322ef41b0159cd83ff9364ea6"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.610625 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4a58c1d-0d71-4853-b894-e14efec97e65" containerID="9ba9996afb713de565f46dd615dc6059f7953e6f9fedcd7a54804fca2ebf7629" exitCode=0 Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.610693 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" event={"ID":"f4a58c1d-0d71-4853-b894-e14efec97e65","Type":"ContainerDied","Data":"9ba9996afb713de565f46dd615dc6059f7953e6f9fedcd7a54804fca2ebf7629"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.610723 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" event={"ID":"f4a58c1d-0d71-4853-b894-e14efec97e65","Type":"ContainerStarted","Data":"476689b4f1f132e2aad51bc2adc918d78b774bd4f2bc57cb2502ec271318d91e"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.611477 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.628413 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" podStartSLOduration=126.628392305 podStartE2EDuration="2m6.628392305s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.622164319 +0000 UTC m=+147.074363576" watchObservedRunningTime="2025-09-30 08:05:23.628392305 +0000 UTC m=+147.080591572" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.642583 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" event={"ID":"626d8738-cb60-49b6-ba04-754600d28a86","Type":"ContainerStarted","Data":"d58b696411401e5a0e7a5644b37ed41d62796ed9de6582b1be32e47d94dfc483"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.657353 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.659744 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p7cc4" podStartSLOduration=125.659719821 podStartE2EDuration="2m5.659719821s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.646690223 +0000 UTC m=+147.098889490" watchObservedRunningTime="2025-09-30 08:05:23.659719821 +0000 UTC m=+147.111919088" Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.660999 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.160982587 +0000 UTC m=+147.613181854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.665155 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" event={"ID":"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c","Type":"ContainerStarted","Data":"d67eba0e824ff1e6c01b95a1666b05fdffb6ed19c767245b1c0fd4011affefa0"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.665237 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" event={"ID":"60f55bdb-8ad2-4e65-b3c2-aed4488e6a1c","Type":"ContainerStarted","Data":"561666ffa5728bf4702a9c7d9065583cb951573b071dea087ba88618995b6189"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.669228 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5q5vw" podStartSLOduration=125.669209039 podStartE2EDuration="2m5.669209039s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.668404576 +0000 UTC m=+147.120603843" watchObservedRunningTime="2025-09-30 08:05:23.669209039 +0000 UTC m=+147.121408306" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.676961 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" event={"ID":"cd62560a-d47e-49f2-b3cd-06568dc86f43","Type":"ContainerStarted","Data":"5e2a599a7237625c5254f081c087d6d6a83bdeacb2c0440003efd146b61573c5"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.685971 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" event={"ID":"6083d85f-fd9a-4df3-84c5-f6a92e84c53c","Type":"ContainerStarted","Data":"f9cac3bfc252e36bfb334487730b3553d714cbd33b831a520352b669d69fec54"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.689474 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" podStartSLOduration=126.689456552 podStartE2EDuration="2m6.689456552s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.687325171 +0000 UTC m=+147.139524438" watchObservedRunningTime="2025-09-30 08:05:23.689456552 +0000 UTC m=+147.141655819" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.698141 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" event={"ID":"c5971693-8b44-4a2a-bd8f-7ee41e0f69cb","Type":"ContainerStarted","Data":"f3e040bfecf3a887ca291690b0257b614176ee95a51c59e911d1d758b9c69c8a"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.698190 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" event={"ID":"c5971693-8b44-4a2a-bd8f-7ee41e0f69cb","Type":"ContainerStarted","Data":"1f5a26a81861d5c20fda0ab080e48add3b07ba18d51b8a66e47a56a38e58cbf5"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.699110 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.702551 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" event={"ID":"5cc8817f-e25c-4a2d-aa65-2696793ff6fc","Type":"ContainerStarted","Data":"330ba1d922aa2c7ff7db84a8f1fb6cea1761bdc6e00d9e62c755c8b3ddd9bcae"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.702585 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" event={"ID":"5cc8817f-e25c-4a2d-aa65-2696793ff6fc","Type":"ContainerStarted","Data":"447d2655dcd7ce6df33b18d04a810f698b3c7a0cd4f5f5a00687477992b97103"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.703105 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.706908 4810 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jrsk2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.706960 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" podUID="5cc8817f-e25c-4a2d-aa65-2696793ff6fc" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.710619 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6lcfr" event={"ID":"1c38bb35-d8db-40f6-903a-10682aa99f35","Type":"ContainerStarted","Data":"29ec348868536060afc32d2ec1cb5b4d0bbc79020caaffde6b7ad262cef44fb4"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.710661 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6lcfr" event={"ID":"1c38bb35-d8db-40f6-903a-10682aa99f35","Type":"ContainerStarted","Data":"ed547dd51d75d538c877bdc930a56728d5982191a8d7ea31f1e99cfaf055af3e"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.711153 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.721938 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-k785t" event={"ID":"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4","Type":"ContainerStarted","Data":"e663a674fa6d6672a0732b3bca46bf12c07c72f39a01f1f0154a3c58fe0d3beb"} Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.723016 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-47kbw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.723057 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-47kbw" podUID="e90bfa19-6003-4593-b318-31ab8ec8c44f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.724133 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.725716 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.729047 4810 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-gnt7r container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.729124 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" podUID="cd62560a-d47e-49f2-b3cd-06568dc86f43" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.752201 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" podStartSLOduration=125.752178505 podStartE2EDuration="2m5.752178505s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.746036141 +0000 UTC m=+147.198235428" watchObservedRunningTime="2025-09-30 08:05:23.752178505 +0000 UTC m=+147.204377782" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.761167 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-49zt7" podStartSLOduration=125.761138388 podStartE2EDuration="2m5.761138388s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.721094616 +0000 UTC m=+147.173293883" watchObservedRunningTime="2025-09-30 08:05:23.761138388 +0000 UTC m=+147.213337675" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.762810 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.764311 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.264289897 +0000 UTC m=+147.716489204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.789694 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-s2vww" podStartSLOduration=126.789672465 podStartE2EDuration="2m6.789672465s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.783001336 +0000 UTC m=+147.235200613" watchObservedRunningTime="2025-09-30 08:05:23.789672465 +0000 UTC m=+147.241871742" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.812978 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtbw5" podStartSLOduration=125.812960023 podStartE2EDuration="2m5.812960023s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.811716998 +0000 UTC m=+147.263916265" watchObservedRunningTime="2025-09-30 08:05:23.812960023 +0000 UTC m=+147.265159280" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.864817 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.867383 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.367368541 +0000 UTC m=+147.819567808 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.887749 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" podStartSLOduration=126.887729677 podStartE2EDuration="2m6.887729677s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.882869669 +0000 UTC m=+147.335068936" watchObservedRunningTime="2025-09-30 08:05:23.887729677 +0000 UTC m=+147.339928944" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.934137 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" podStartSLOduration=126.934120378 podStartE2EDuration="2m6.934120378s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.931688209 +0000 UTC m=+147.383887476" watchObservedRunningTime="2025-09-30 08:05:23.934120378 +0000 UTC m=+147.386319645" Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.965930 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.966045 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.46602077 +0000 UTC m=+147.918220037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.966286 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:23 crc kubenswrapper[4810]: E0930 08:05:23.966713 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.466694629 +0000 UTC m=+147.918893896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:23 crc kubenswrapper[4810]: I0930 08:05:23.997926 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" podStartSLOduration=125.997908721 podStartE2EDuration="2m5.997908721s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:23.996995285 +0000 UTC m=+147.449194552" watchObservedRunningTime="2025-09-30 08:05:23.997908721 +0000 UTC m=+147.450107988" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.013978 4810 patch_prober.go:28] interesting pod/router-default-5444994796-555gv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 08:05:24 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Sep 30 08:05:24 crc kubenswrapper[4810]: [+]process-running ok Sep 30 08:05:24 crc kubenswrapper[4810]: healthz check failed Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.014042 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-555gv" podUID="f4b41c70-05a1-4342-8aee-cba6e44b89dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.050497 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" podStartSLOduration=126.050480627 podStartE2EDuration="2m6.050480627s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:24.044118918 +0000 UTC m=+147.496318185" watchObservedRunningTime="2025-09-30 08:05:24.050480627 +0000 UTC m=+147.502679894" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.066937 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.067123 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.567092547 +0000 UTC m=+148.019291814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.067246 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.067541 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.56753406 +0000 UTC m=+148.019733327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.077222 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rh9mj" podStartSLOduration=127.077204573 podStartE2EDuration="2m7.077204573s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:24.075163175 +0000 UTC m=+147.527362442" watchObservedRunningTime="2025-09-30 08:05:24.077204573 +0000 UTC m=+147.529403840" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.103255 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5wczx" podStartSLOduration=127.103237729 podStartE2EDuration="2m7.103237729s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:24.101349465 +0000 UTC m=+147.553548742" watchObservedRunningTime="2025-09-30 08:05:24.103237729 +0000 UTC m=+147.555436996" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.153210 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-6lcfr" podStartSLOduration=8.153190841 podStartE2EDuration="8.153190841s" podCreationTimestamp="2025-09-30 08:05:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:24.132417654 +0000 UTC m=+147.584616921" watchObservedRunningTime="2025-09-30 08:05:24.153190841 +0000 UTC m=+147.605390108" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.168727 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.169137 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.669109961 +0000 UTC m=+148.121309228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.169430 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.169812 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.66979583 +0000 UTC m=+148.121995097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.170733 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" podStartSLOduration=126.170712706 podStartE2EDuration="2m6.170712706s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:24.169496972 +0000 UTC m=+147.621696239" watchObservedRunningTime="2025-09-30 08:05:24.170712706 +0000 UTC m=+147.622911973" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.201484 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" podStartSLOduration=126.201469196 podStartE2EDuration="2m6.201469196s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:24.200389955 +0000 UTC m=+147.652589222" watchObservedRunningTime="2025-09-30 08:05:24.201469196 +0000 UTC m=+147.653668463" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.270197 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.270433 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.770403405 +0000 UTC m=+148.222602672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.270528 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.270800 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.770787595 +0000 UTC m=+148.222986862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.372336 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.372555 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.872525811 +0000 UTC m=+148.324725078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.372703 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.373004 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.872988945 +0000 UTC m=+148.325188212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.474279 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.474501 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.974474533 +0000 UTC m=+148.426673790 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.474556 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.474993 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:24.974971848 +0000 UTC m=+148.427171135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.576167 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.576381 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.076355164 +0000 UTC m=+148.528554421 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.576472 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.576816 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.076803186 +0000 UTC m=+148.529002443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.677421 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.677602 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.177566765 +0000 UTC m=+148.629766022 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.677700 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.678018 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.178005347 +0000 UTC m=+148.630204614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.729595 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" event={"ID":"6760d07c-3b94-492f-b58c-c61dbd1df867","Type":"ContainerStarted","Data":"e082186d0446f46d483c30a1570c44a1f8ad19bce8919261a844926d27d8ca87"} Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.731020 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6gc9d" event={"ID":"06d3ea99-b8e6-4e46-bd04-e370f73274ef","Type":"ContainerStarted","Data":"1af46bd17e1998fd2b278fe3cbe4f0df3c069a240a5dbe0a606b18ad6704b49c"} Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.733346 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6lcfr" event={"ID":"1c38bb35-d8db-40f6-903a-10682aa99f35","Type":"ContainerStarted","Data":"e6bce3dc0d31d30c0b4c04bf02f7c0c6a544728be88758582fe274e406d37c67"} Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.735034 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-k785t" event={"ID":"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4","Type":"ContainerStarted","Data":"c8e41603bd1bbb484514ab65f1962050404727a0747fc394c84df64bc657633b"} Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.739337 4810 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tmk5l container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.739396 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" podUID="e049bda6-e38f-4357-926e-d8b767cf5ed9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.753153 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5h6v8" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.762600 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfd6l" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.778872 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.779086 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.279060114 +0000 UTC m=+148.731259381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.779297 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.779667 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.279648201 +0000 UTC m=+148.731847468 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.800962 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-5lb54" podStartSLOduration=126.800946593 podStartE2EDuration="2m6.800946593s" podCreationTimestamp="2025-09-30 08:03:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:24.772386565 +0000 UTC m=+148.224585832" watchObservedRunningTime="2025-09-30 08:05:24.800946593 +0000 UTC m=+148.253145860" Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.880958 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.881150 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.381121509 +0000 UTC m=+148.833320776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.881921 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.884423 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.384406622 +0000 UTC m=+148.836605889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.983623 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.983792 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.483762751 +0000 UTC m=+148.935962018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:24 crc kubenswrapper[4810]: I0930 08:05:24.983965 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:24 crc kubenswrapper[4810]: E0930 08:05:24.984333 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.484325497 +0000 UTC m=+148.936524754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.017762 4810 patch_prober.go:28] interesting pod/router-default-5444994796-555gv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 08:05:25 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Sep 30 08:05:25 crc kubenswrapper[4810]: [+]process-running ok Sep 30 08:05:25 crc kubenswrapper[4810]: healthz check failed Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.017828 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-555gv" podUID="f4b41c70-05a1-4342-8aee-cba6e44b89dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.084647 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.084864 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.584838208 +0000 UTC m=+149.037037475 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.085026 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.085355 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.585341333 +0000 UTC m=+149.037540600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.186046 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.186238 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.686208624 +0000 UTC m=+149.138407891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.186853 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.187294 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.687277154 +0000 UTC m=+149.139476421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.287617 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.287789 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.287846 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.787817967 +0000 UTC m=+149.240017234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.287901 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.287961 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.288042 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.288104 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.288422 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.788403633 +0000 UTC m=+149.240602900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.288802 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.297146 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.299892 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.300429 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.328694 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.329555 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.344324 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.390778 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.391045 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:25.891030274 +0000 UTC m=+149.343229541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.449476 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.494916 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.516176 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:26.01614697 +0000 UTC m=+149.468346237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.582625 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jrsk2" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.608655 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.609037 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:26.109019956 +0000 UTC m=+149.561219223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.694256 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6vwd5"] Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.695347 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.701191 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.718817 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.719136 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:26.219123468 +0000 UTC m=+149.671322735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.736345 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6vwd5"] Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.783619 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-k785t" event={"ID":"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4","Type":"ContainerStarted","Data":"cba3646ffe5c662356b2e77d257bf9abcc1e0965654478f7de46d4fd00cc2f26"} Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.783944 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-k785t" event={"ID":"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4","Type":"ContainerStarted","Data":"e6dec5e7ee7e66e23e6da35ae099b615dcef39a443180391062fb5099533fbbf"} Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.784979 4810 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tmk5l container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.785022 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" podUID="e049bda6-e38f-4357-926e-d8b767cf5ed9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.826328 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.826574 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-catalog-content\") pod \"certified-operators-6vwd5\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.826604 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-utilities\") pod \"certified-operators-6vwd5\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.826646 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts6qk\" (UniqueName: \"kubernetes.io/projected/015f2a85-90c3-4bd9-b77a-63209e20cc74-kube-api-access-ts6qk\") pod \"certified-operators-6vwd5\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.826778 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:26.326762431 +0000 UTC m=+149.778961698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.870710 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t8pf8"] Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.874705 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.878807 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.928500 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-catalog-content\") pod \"certified-operators-6vwd5\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.928608 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-utilities\") pod \"certified-operators-6vwd5\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.928779 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts6qk\" (UniqueName: \"kubernetes.io/projected/015f2a85-90c3-4bd9-b77a-63209e20cc74-kube-api-access-ts6qk\") pod \"certified-operators-6vwd5\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.928821 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.933143 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-catalog-content\") pod \"certified-operators-6vwd5\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.938939 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-utilities\") pod \"certified-operators-6vwd5\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:05:25 crc kubenswrapper[4810]: E0930 08:05:25.939677 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:26.439659213 +0000 UTC m=+149.891858470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.980422 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t8pf8"] Sep 30 08:05:25 crc kubenswrapper[4810]: I0930 08:05:25.983752 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts6qk\" (UniqueName: \"kubernetes.io/projected/015f2a85-90c3-4bd9-b77a-63209e20cc74-kube-api-access-ts6qk\") pod \"certified-operators-6vwd5\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.022247 4810 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.026573 4810 patch_prober.go:28] interesting pod/router-default-5444994796-555gv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 08:05:26 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Sep 30 08:05:26 crc kubenswrapper[4810]: [+]process-running ok Sep 30 08:05:26 crc kubenswrapper[4810]: healthz check failed Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.026634 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-555gv" podUID="f4b41c70-05a1-4342-8aee-cba6e44b89dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.033032 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.033251 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-utilities\") pod \"community-operators-t8pf8\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.033307 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-catalog-content\") pod \"community-operators-t8pf8\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.033374 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxhpx\" (UniqueName: \"kubernetes.io/projected/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-kube-api-access-rxhpx\") pod \"community-operators-t8pf8\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:05:26 crc kubenswrapper[4810]: E0930 08:05:26.033468 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:26.533451925 +0000 UTC m=+149.985651192 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.057619 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.082509 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m27w2"] Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.083771 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.093315 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m27w2"] Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.137040 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwnn8\" (UniqueName: \"kubernetes.io/projected/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-kube-api-access-jwnn8\") pod \"certified-operators-m27w2\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.137106 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxhpx\" (UniqueName: \"kubernetes.io/projected/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-kube-api-access-rxhpx\") pod \"community-operators-t8pf8\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.137142 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-catalog-content\") pod \"certified-operators-m27w2\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.137168 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.137189 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-utilities\") pod \"certified-operators-m27w2\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.137206 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-utilities\") pod \"community-operators-t8pf8\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.137230 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-catalog-content\") pod \"community-operators-t8pf8\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.137666 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-catalog-content\") pod \"community-operators-t8pf8\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:05:26 crc kubenswrapper[4810]: E0930 08:05:26.138205 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:26.638194696 +0000 UTC m=+150.090393963 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.138441 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-utilities\") pod \"community-operators-t8pf8\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.157363 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxhpx\" (UniqueName: \"kubernetes.io/projected/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-kube-api-access-rxhpx\") pod \"community-operators-t8pf8\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.237825 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.238299 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwnn8\" (UniqueName: \"kubernetes.io/projected/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-kube-api-access-jwnn8\") pod \"certified-operators-m27w2\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.238346 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-catalog-content\") pod \"certified-operators-m27w2\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.238385 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-utilities\") pod \"certified-operators-m27w2\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.238830 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-utilities\") pod \"certified-operators-m27w2\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:05:26 crc kubenswrapper[4810]: E0930 08:05:26.238895 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:26.738880692 +0000 UTC m=+150.191079959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.243345 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-catalog-content\") pod \"certified-operators-m27w2\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.267931 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwnn8\" (UniqueName: \"kubernetes.io/projected/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-kube-api-access-jwnn8\") pod \"certified-operators-m27w2\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.276707 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lrxfp"] Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.277653 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.280304 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.280342 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lrxfp"] Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.339452 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.339946 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-utilities\") pod \"community-operators-lrxfp\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.339977 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-catalog-content\") pod \"community-operators-lrxfp\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.340037 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls599\" (UniqueName: \"kubernetes.io/projected/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-kube-api-access-ls599\") pod \"community-operators-lrxfp\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:05:26 crc kubenswrapper[4810]: E0930 08:05:26.340504 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 08:05:26.840484984 +0000 UTC m=+150.292684251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4bvx" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.434207 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.440885 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.441097 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-utilities\") pod \"community-operators-lrxfp\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.441135 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-catalog-content\") pod \"community-operators-lrxfp\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.441166 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls599\" (UniqueName: \"kubernetes.io/projected/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-kube-api-access-ls599\") pod \"community-operators-lrxfp\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:05:26 crc kubenswrapper[4810]: E0930 08:05:26.441640 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 08:05:26.941624954 +0000 UTC m=+150.393824221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.500830 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6vwd5"] Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.502375 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-utilities\") pod \"community-operators-lrxfp\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.502487 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-catalog-content\") pod \"community-operators-lrxfp\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.504239 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls599\" (UniqueName: \"kubernetes.io/projected/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-kube-api-access-ls599\") pod \"community-operators-lrxfp\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.516401 4810 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-09-30T08:05:26.022283329Z","Handler":null,"Name":""} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.524528 4810 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.524562 4810 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.542436 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.550360 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.550412 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.591967 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4bvx\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.643793 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.645667 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t8pf8"] Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.662529 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 30 08:05:26 crc kubenswrapper[4810]: W0930 08:05:26.662828 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51ccfc9f_81fc_4c77_95bb_fdb5e2239829.slice/crio-f26133c9396f16dbf2dc05d741d26ed1819de76354146e5981d587c965a0e271 WatchSource:0}: Error finding container f26133c9396f16dbf2dc05d741d26ed1819de76354146e5981d587c965a0e271: Status 404 returned error can't find the container with id f26133c9396f16dbf2dc05d741d26ed1819de76354146e5981d587c965a0e271 Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.684905 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.747937 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m27w2"] Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.789101 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.804066 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"76c7313ef7923ba717fb838da314176473c6c82f591e06b085ded1db1e290d0c"} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.804099 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"458625f43beeede1c75940b640da4ba61363a4fbe761c32aa2b580cdf44090d2"} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.805220 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6ea3990a73e0dcd0eb81070e16b0e0102dfd9df9d36f7b17b10aa7a85bcc2ecd"} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.805437 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"894e981e605a4c465bd3d5e708d8162fefb78b47b69901a7350f48c909ee6c8e"} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.808467 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8pf8" event={"ID":"51ccfc9f-81fc-4c77-95bb-fdb5e2239829","Type":"ContainerStarted","Data":"f26133c9396f16dbf2dc05d741d26ed1819de76354146e5981d587c965a0e271"} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.814102 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-k785t" event={"ID":"e7bf6704-7c12-4f9d-b0bf-e1592fae2ec4","Type":"ContainerStarted","Data":"605c3818a82101b2fb8aea3f248e9b0d1f5668c5b2f024f52e682eb2d8925afe"} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.823210 4810 generic.go:334] "Generic (PLEG): container finished" podID="015f2a85-90c3-4bd9-b77a-63209e20cc74" containerID="3a66a215df1ea740c0c89f0c95bde8f7cb9d2acb79d8161a6eb47f8ffc1d2c1e" exitCode=0 Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.823373 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vwd5" event={"ID":"015f2a85-90c3-4bd9-b77a-63209e20cc74","Type":"ContainerDied","Data":"3a66a215df1ea740c0c89f0c95bde8f7cb9d2acb79d8161a6eb47f8ffc1d2c1e"} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.823409 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vwd5" event={"ID":"015f2a85-90c3-4bd9-b77a-63209e20cc74","Type":"ContainerStarted","Data":"45214645d40e376161531e77af5d98ccfaa2c3832f6a836c2f6c4e0e047df397"} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.830885 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"96f6dc1c659c09a96e74fd03636495a5cd3890166b9f7b011899dac4b741069c"} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.830936 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7e20847ca70744a0b0aa5c7db9883d5e350ef5b0ae61c14b1b3a250c82669a4c"} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.831544 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.833922 4810 generic.go:334] "Generic (PLEG): container finished" podID="b8d5914f-a41d-4e29-bd6b-3e5050c23cd6" containerID="26b716174b8617ef9307046c3f0d04ab2fdc57b54a9b7691dae85ca34cd14319" exitCode=0 Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.834034 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" event={"ID":"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6","Type":"ContainerDied","Data":"26b716174b8617ef9307046c3f0d04ab2fdc57b54a9b7691dae85ca34cd14319"} Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.845284 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.858473 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-k785t" podStartSLOduration=10.858455958 podStartE2EDuration="10.858455958s" podCreationTimestamp="2025-09-30 08:05:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:26.856122022 +0000 UTC m=+150.308321289" watchObservedRunningTime="2025-09-30 08:05:26.858455958 +0000 UTC m=+150.310655225" Sep 30 08:05:26 crc kubenswrapper[4810]: I0930 08:05:26.942604 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lrxfp"] Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.013775 4810 patch_prober.go:28] interesting pod/router-default-5444994796-555gv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 08:05:27 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Sep 30 08:05:27 crc kubenswrapper[4810]: [+]process-running ok Sep 30 08:05:27 crc kubenswrapper[4810]: healthz check failed Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.014175 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-555gv" podUID="f4b41c70-05a1-4342-8aee-cba6e44b89dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.081759 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4bvx"] Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.336191 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.842347 4810 generic.go:334] "Generic (PLEG): container finished" podID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" containerID="359e775036370e0921095f82a9f45a3b857f47b99e122aadb3991e995d4f48b2" exitCode=0 Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.842469 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrxfp" event={"ID":"76f9c24e-1aa7-4751-84c5-1a67afcbf75b","Type":"ContainerDied","Data":"359e775036370e0921095f82a9f45a3b857f47b99e122aadb3991e995d4f48b2"} Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.842647 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrxfp" event={"ID":"76f9c24e-1aa7-4751-84c5-1a67afcbf75b","Type":"ContainerStarted","Data":"22c7ba6e8b7bb503cb2b84fdfe598ddcea2b64ff0e8b96527893747e9ac3987d"} Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.845672 4810 generic.go:334] "Generic (PLEG): container finished" podID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" containerID="260043361c1ef49108c4043fd614f87158b44615189c735fe31bcf08aa862cb6" exitCode=0 Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.845721 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8pf8" event={"ID":"51ccfc9f-81fc-4c77-95bb-fdb5e2239829","Type":"ContainerDied","Data":"260043361c1ef49108c4043fd614f87158b44615189c735fe31bcf08aa862cb6"} Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.866435 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqwr"] Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.867731 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" event={"ID":"4c2e4c71-7e20-4d81-8531-f673f4c9341f","Type":"ContainerStarted","Data":"6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6"} Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.867765 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.867779 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" event={"ID":"4c2e4c71-7e20-4d81-8531-f673f4c9341f","Type":"ContainerStarted","Data":"3ec7bd439fc03cf75c8da3e4b679811544dc48a87d854c93c7348939eaac5acd"} Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.867895 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.870255 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.874405 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqwr"] Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.889305 4810 generic.go:334] "Generic (PLEG): container finished" podID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" containerID="5e3e1081aad279c411b4eb6b9d912295a83fc4bbe0d753b8c6b85de0b7ba413f" exitCode=0 Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.889637 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m27w2" event={"ID":"cb5387a0-ff9e-42bc-85b4-16e50a6f19de","Type":"ContainerDied","Data":"5e3e1081aad279c411b4eb6b9d912295a83fc4bbe0d753b8c6b85de0b7ba413f"} Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.889706 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m27w2" event={"ID":"cb5387a0-ff9e-42bc-85b4-16e50a6f19de","Type":"ContainerStarted","Data":"1ffe784c60f5532a1d867aec8d4c207e0c71f2d8d98dcb20f10d71a07ce5bb42"} Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.934237 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" podStartSLOduration=130.934211709 podStartE2EDuration="2m10.934211709s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:27.908059719 +0000 UTC m=+151.360258986" watchObservedRunningTime="2025-09-30 08:05:27.934211709 +0000 UTC m=+151.386410976" Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.969294 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-catalog-content\") pod \"redhat-marketplace-xcqwr\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.969485 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-utilities\") pod \"redhat-marketplace-xcqwr\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:05:27 crc kubenswrapper[4810]: I0930 08:05:27.969505 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m6bg\" (UniqueName: \"kubernetes.io/projected/fb1baa89-d430-4243-9b2d-2d2932c1abe4-kube-api-access-4m6bg\") pod \"redhat-marketplace-xcqwr\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.014811 4810 patch_prober.go:28] interesting pod/router-default-5444994796-555gv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 08:05:28 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Sep 30 08:05:28 crc kubenswrapper[4810]: [+]process-running ok Sep 30 08:05:28 crc kubenswrapper[4810]: healthz check failed Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.014890 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-555gv" podUID="f4b41c70-05a1-4342-8aee-cba6e44b89dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.070801 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-catalog-content\") pod \"redhat-marketplace-xcqwr\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.070921 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-utilities\") pod \"redhat-marketplace-xcqwr\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.070941 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m6bg\" (UniqueName: \"kubernetes.io/projected/fb1baa89-d430-4243-9b2d-2d2932c1abe4-kube-api-access-4m6bg\") pod \"redhat-marketplace-xcqwr\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.071487 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-catalog-content\") pod \"redhat-marketplace-xcqwr\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.071694 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-utilities\") pod \"redhat-marketplace-xcqwr\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.123804 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m6bg\" (UniqueName: \"kubernetes.io/projected/fb1baa89-d430-4243-9b2d-2d2932c1abe4-kube-api-access-4m6bg\") pod \"redhat-marketplace-xcqwr\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.182341 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.189760 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.252504 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nqdvv"] Sep 30 08:05:28 crc kubenswrapper[4810]: E0930 08:05:28.252779 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8d5914f-a41d-4e29-bd6b-3e5050c23cd6" containerName="collect-profiles" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.252795 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8d5914f-a41d-4e29-bd6b-3e5050c23cd6" containerName="collect-profiles" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.252888 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8d5914f-a41d-4e29-bd6b-3e5050c23cd6" containerName="collect-profiles" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.253612 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.262970 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqdvv"] Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.273082 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-secret-volume\") pod \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.273168 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-config-volume\") pod \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.273214 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sck96\" (UniqueName: \"kubernetes.io/projected/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-kube-api-access-sck96\") pod \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\" (UID: \"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6\") " Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.273858 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-config-volume" (OuterVolumeSpecName: "config-volume") pod "b8d5914f-a41d-4e29-bd6b-3e5050c23cd6" (UID: "b8d5914f-a41d-4e29-bd6b-3e5050c23cd6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.276365 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b8d5914f-a41d-4e29-bd6b-3e5050c23cd6" (UID: "b8d5914f-a41d-4e29-bd6b-3e5050c23cd6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.280526 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-kube-api-access-sck96" (OuterVolumeSpecName: "kube-api-access-sck96") pod "b8d5914f-a41d-4e29-bd6b-3e5050c23cd6" (UID: "b8d5914f-a41d-4e29-bd6b-3e5050c23cd6"). InnerVolumeSpecName "kube-api-access-sck96". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.349478 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.350464 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.361570 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.361815 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.382370 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.384297 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/48ddf04a-273c-491e-a152-a67ef31eb463-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"48ddf04a-273c-491e-a152-a67ef31eb463\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.384362 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjl8c\" (UniqueName: \"kubernetes.io/projected/8e6d85f7-29e2-41a8-b65e-02361a42f81d-kube-api-access-wjl8c\") pod \"redhat-marketplace-nqdvv\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.384411 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48ddf04a-273c-491e-a152-a67ef31eb463-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"48ddf04a-273c-491e-a152-a67ef31eb463\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.384437 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-utilities\") pod \"redhat-marketplace-nqdvv\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.384503 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-catalog-content\") pod \"redhat-marketplace-nqdvv\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.384554 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.384569 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sck96\" (UniqueName: \"kubernetes.io/projected/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-kube-api-access-sck96\") on node \"crc\" DevicePath \"\"" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.384584 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.485175 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-catalog-content\") pod \"redhat-marketplace-nqdvv\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.485237 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/48ddf04a-273c-491e-a152-a67ef31eb463-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"48ddf04a-273c-491e-a152-a67ef31eb463\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.485345 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjl8c\" (UniqueName: \"kubernetes.io/projected/8e6d85f7-29e2-41a8-b65e-02361a42f81d-kube-api-access-wjl8c\") pod \"redhat-marketplace-nqdvv\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.485389 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48ddf04a-273c-491e-a152-a67ef31eb463-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"48ddf04a-273c-491e-a152-a67ef31eb463\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.485415 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-utilities\") pod \"redhat-marketplace-nqdvv\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.485817 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/48ddf04a-273c-491e-a152-a67ef31eb463-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"48ddf04a-273c-491e-a152-a67ef31eb463\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.485852 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-utilities\") pod \"redhat-marketplace-nqdvv\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.486167 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-catalog-content\") pod \"redhat-marketplace-nqdvv\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.503945 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjl8c\" (UniqueName: \"kubernetes.io/projected/8e6d85f7-29e2-41a8-b65e-02361a42f81d-kube-api-access-wjl8c\") pod \"redhat-marketplace-nqdvv\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.508150 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqwr"] Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.513314 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48ddf04a-273c-491e-a152-a67ef31eb463-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"48ddf04a-273c-491e-a152-a67ef31eb463\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.514025 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.514187 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.523866 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.581733 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.602432 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rg6tm" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.696733 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.734610 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.744512 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gnt7r" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.875494 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8x56x"] Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.882736 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.890475 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.894786 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8x56x"] Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.936804 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" event={"ID":"b8d5914f-a41d-4e29-bd6b-3e5050c23cd6","Type":"ContainerDied","Data":"cc7789b59dc790bee1bc7496907de5f4e19da5871b997c4f48aa89d0408b797e"} Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.937113 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc7789b59dc790bee1bc7496907de5f4e19da5871b997c4f48aa89d0408b797e" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.937241 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn" Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.972070 4810 generic.go:334] "Generic (PLEG): container finished" podID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" containerID="32c465e63ce961f8d3c7edd7f41e046e00779a7ce4f911f0b001f2c539f23d33" exitCode=0 Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.972183 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqwr" event={"ID":"fb1baa89-d430-4243-9b2d-2d2932c1abe4","Type":"ContainerDied","Data":"32c465e63ce961f8d3c7edd7f41e046e00779a7ce4f911f0b001f2c539f23d33"} Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.972210 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqwr" event={"ID":"fb1baa89-d430-4243-9b2d-2d2932c1abe4","Type":"ContainerStarted","Data":"cd78dd2ea82a1fb37445b03443d83e09fda67ba2c2e4eb926f431c84fb4418f4"} Sep 30 08:05:28 crc kubenswrapper[4810]: I0930 08:05:28.988157 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-pq7h8" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.004016 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-catalog-content\") pod \"redhat-operators-8x56x\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.004088 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-utilities\") pod \"redhat-operators-8x56x\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.004128 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b49p\" (UniqueName: \"kubernetes.io/projected/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-kube-api-access-2b49p\") pod \"redhat-operators-8x56x\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.012380 4810 patch_prober.go:28] interesting pod/router-default-5444994796-555gv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 08:05:29 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Sep 30 08:05:29 crc kubenswrapper[4810]: [+]process-running ok Sep 30 08:05:29 crc kubenswrapper[4810]: healthz check failed Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.012432 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-555gv" podUID="f4b41c70-05a1-4342-8aee-cba6e44b89dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.056859 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.058193 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.069052 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqdvv"] Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.079800 4810 patch_prober.go:28] interesting pod/console-f9d7485db-5w577 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.079861 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5w577" podUID="64fd38ac-d0ef-430c-aed5-5a4e4944986c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.106895 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-utilities\") pod \"redhat-operators-8x56x\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.107001 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b49p\" (UniqueName: \"kubernetes.io/projected/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-kube-api-access-2b49p\") pod \"redhat-operators-8x56x\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.107196 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-catalog-content\") pod \"redhat-operators-8x56x\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.110338 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-catalog-content\") pod \"redhat-operators-8x56x\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.115576 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-utilities\") pod \"redhat-operators-8x56x\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.144855 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b49p\" (UniqueName: \"kubernetes.io/projected/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-kube-api-access-2b49p\") pod \"redhat-operators-8x56x\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.230773 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.263309 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wprmn"] Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.264888 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.272204 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wprmn"] Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.308998 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brpsd\" (UniqueName: \"kubernetes.io/projected/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-kube-api-access-brpsd\") pod \"redhat-operators-wprmn\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.309048 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-utilities\") pod \"redhat-operators-wprmn\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.309073 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-catalog-content\") pod \"redhat-operators-wprmn\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.405040 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.412370 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brpsd\" (UniqueName: \"kubernetes.io/projected/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-kube-api-access-brpsd\") pod \"redhat-operators-wprmn\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.412401 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-utilities\") pod \"redhat-operators-wprmn\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.412579 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-catalog-content\") pod \"redhat-operators-wprmn\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.413023 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-catalog-content\") pod \"redhat-operators-wprmn\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.413230 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-utilities\") pod \"redhat-operators-wprmn\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.433878 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brpsd\" (UniqueName: \"kubernetes.io/projected/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-kube-api-access-brpsd\") pod \"redhat-operators-wprmn\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.571035 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8x56x"] Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.574913 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-47kbw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.575419 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-47kbw" podUID="e90bfa19-6003-4593-b318-31ab8ec8c44f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.576815 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-47kbw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.576841 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-47kbw" podUID="e90bfa19-6003-4593-b318-31ab8ec8c44f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Sep 30 08:05:29 crc kubenswrapper[4810]: I0930 08:05:29.594549 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:05:30 crc kubenswrapper[4810]: I0930 08:05:30.001113 4810 generic.go:334] "Generic (PLEG): container finished" podID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" containerID="ec0ac67a412f96254510f299f8521ef3a4d88ce06cea4933c7db87b5d7400ec8" exitCode=0 Sep 30 08:05:30 crc kubenswrapper[4810]: I0930 08:05:30.001248 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqdvv" event={"ID":"8e6d85f7-29e2-41a8-b65e-02361a42f81d","Type":"ContainerDied","Data":"ec0ac67a412f96254510f299f8521ef3a4d88ce06cea4933c7db87b5d7400ec8"} Sep 30 08:05:30 crc kubenswrapper[4810]: I0930 08:05:30.001838 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqdvv" event={"ID":"8e6d85f7-29e2-41a8-b65e-02361a42f81d","Type":"ContainerStarted","Data":"b47398b5e4aedb7b1fe6f78fded5bd0ae58552e56bea54613269bbe023ff68d5"} Sep 30 08:05:30 crc kubenswrapper[4810]: I0930 08:05:30.003851 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"48ddf04a-273c-491e-a152-a67ef31eb463","Type":"ContainerStarted","Data":"8f93a9758ca9a4c0a67fa556380c6b269c970283bdd0a66ab8e032d8a6f4b1bc"} Sep 30 08:05:30 crc kubenswrapper[4810]: I0930 08:05:30.010929 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:30 crc kubenswrapper[4810]: I0930 08:05:30.023410 4810 patch_prober.go:28] interesting pod/router-default-5444994796-555gv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 08:05:30 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Sep 30 08:05:30 crc kubenswrapper[4810]: [+]process-running ok Sep 30 08:05:30 crc kubenswrapper[4810]: healthz check failed Sep 30 08:05:30 crc kubenswrapper[4810]: I0930 08:05:30.023470 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-555gv" podUID="f4b41c70-05a1-4342-8aee-cba6e44b89dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 08:05:30 crc kubenswrapper[4810]: I0930 08:05:30.026278 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8x56x" event={"ID":"fee7bb01-669e-49de-b6fa-7cf3951fdfc2","Type":"ContainerStarted","Data":"11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029"} Sep 30 08:05:30 crc kubenswrapper[4810]: I0930 08:05:30.026307 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8x56x" event={"ID":"fee7bb01-669e-49de-b6fa-7cf3951fdfc2","Type":"ContainerStarted","Data":"3e194d210b9ab7b15870ed46eab69bd4d7f124cb4e0d52d13d15068d56f58741"} Sep 30 08:05:30 crc kubenswrapper[4810]: I0930 08:05:30.054473 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wprmn"] Sep 30 08:05:30 crc kubenswrapper[4810]: W0930 08:05:30.108829 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29a5cb45_7f5b_4700_b80c_3c60a0118d8c.slice/crio-a4d70b622f1a0f2add55fe204b1bb85c88b7188795e1d1ec0c4d668fb813cab5 WatchSource:0}: Error finding container a4d70b622f1a0f2add55fe204b1bb85c88b7188795e1d1ec0c4d668fb813cab5: Status 404 returned error can't find the container with id a4d70b622f1a0f2add55fe204b1bb85c88b7188795e1d1ec0c4d668fb813cab5 Sep 30 08:05:30 crc kubenswrapper[4810]: I0930 08:05:30.120410 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:05:31 crc kubenswrapper[4810]: I0930 08:05:31.016604 4810 patch_prober.go:28] interesting pod/router-default-5444994796-555gv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 08:05:31 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Sep 30 08:05:31 crc kubenswrapper[4810]: [+]process-running ok Sep 30 08:05:31 crc kubenswrapper[4810]: healthz check failed Sep 30 08:05:31 crc kubenswrapper[4810]: I0930 08:05:31.016898 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-555gv" podUID="f4b41c70-05a1-4342-8aee-cba6e44b89dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 08:05:31 crc kubenswrapper[4810]: I0930 08:05:31.057543 4810 generic.go:334] "Generic (PLEG): container finished" podID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" containerID="f0f06aeb11843e40a0f90e83f436a5a9c736e44ae35f5cb92ee45e57b41b3fca" exitCode=0 Sep 30 08:05:31 crc kubenswrapper[4810]: I0930 08:05:31.057615 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wprmn" event={"ID":"29a5cb45-7f5b-4700-b80c-3c60a0118d8c","Type":"ContainerDied","Data":"f0f06aeb11843e40a0f90e83f436a5a9c736e44ae35f5cb92ee45e57b41b3fca"} Sep 30 08:05:31 crc kubenswrapper[4810]: I0930 08:05:31.057642 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wprmn" event={"ID":"29a5cb45-7f5b-4700-b80c-3c60a0118d8c","Type":"ContainerStarted","Data":"a4d70b622f1a0f2add55fe204b1bb85c88b7188795e1d1ec0c4d668fb813cab5"} Sep 30 08:05:31 crc kubenswrapper[4810]: I0930 08:05:31.079224 4810 generic.go:334] "Generic (PLEG): container finished" podID="48ddf04a-273c-491e-a152-a67ef31eb463" containerID="2f0f3ed4c2e8c09a05c4a70de5a85927dc08517b4d73d7a06a0f4debf4860809" exitCode=0 Sep 30 08:05:31 crc kubenswrapper[4810]: I0930 08:05:31.079312 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"48ddf04a-273c-491e-a152-a67ef31eb463","Type":"ContainerDied","Data":"2f0f3ed4c2e8c09a05c4a70de5a85927dc08517b4d73d7a06a0f4debf4860809"} Sep 30 08:05:31 crc kubenswrapper[4810]: I0930 08:05:31.087797 4810 generic.go:334] "Generic (PLEG): container finished" podID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" containerID="11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029" exitCode=0 Sep 30 08:05:31 crc kubenswrapper[4810]: I0930 08:05:31.087834 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8x56x" event={"ID":"fee7bb01-669e-49de-b6fa-7cf3951fdfc2","Type":"ContainerDied","Data":"11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029"} Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.011450 4810 patch_prober.go:28] interesting pod/router-default-5444994796-555gv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 08:05:32 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Sep 30 08:05:32 crc kubenswrapper[4810]: [+]process-running ok Sep 30 08:05:32 crc kubenswrapper[4810]: healthz check failed Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.011595 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-555gv" podUID="f4b41c70-05a1-4342-8aee-cba6e44b89dc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.305070 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.305923 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.308545 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.308788 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.357093 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.390004 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3364826a-a3b8-43ec-b28a-4e616f064cf4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3364826a-a3b8-43ec-b28a-4e616f064cf4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.390068 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3364826a-a3b8-43ec-b28a-4e616f064cf4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3364826a-a3b8-43ec-b28a-4e616f064cf4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.490847 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3364826a-a3b8-43ec-b28a-4e616f064cf4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3364826a-a3b8-43ec-b28a-4e616f064cf4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.490902 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3364826a-a3b8-43ec-b28a-4e616f064cf4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3364826a-a3b8-43ec-b28a-4e616f064cf4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.491291 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3364826a-a3b8-43ec-b28a-4e616f064cf4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3364826a-a3b8-43ec-b28a-4e616f064cf4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.513044 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3364826a-a3b8-43ec-b28a-4e616f064cf4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3364826a-a3b8-43ec-b28a-4e616f064cf4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.562685 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.640978 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.700061 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/48ddf04a-273c-491e-a152-a67ef31eb463-kubelet-dir\") pod \"48ddf04a-273c-491e-a152-a67ef31eb463\" (UID: \"48ddf04a-273c-491e-a152-a67ef31eb463\") " Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.700162 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48ddf04a-273c-491e-a152-a67ef31eb463-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "48ddf04a-273c-491e-a152-a67ef31eb463" (UID: "48ddf04a-273c-491e-a152-a67ef31eb463"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.700232 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48ddf04a-273c-491e-a152-a67ef31eb463-kube-api-access\") pod \"48ddf04a-273c-491e-a152-a67ef31eb463\" (UID: \"48ddf04a-273c-491e-a152-a67ef31eb463\") " Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.700637 4810 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/48ddf04a-273c-491e-a152-a67ef31eb463-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.704577 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48ddf04a-273c-491e-a152-a67ef31eb463-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "48ddf04a-273c-491e-a152-a67ef31eb463" (UID: "48ddf04a-273c-491e-a152-a67ef31eb463"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.802656 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48ddf04a-273c-491e-a152-a67ef31eb463-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 08:05:32 crc kubenswrapper[4810]: I0930 08:05:32.960165 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 08:05:33 crc kubenswrapper[4810]: I0930 08:05:33.012687 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:33 crc kubenswrapper[4810]: I0930 08:05:33.015596 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-555gv" Sep 30 08:05:33 crc kubenswrapper[4810]: W0930 08:05:33.047750 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3364826a_a3b8_43ec_b28a_4e616f064cf4.slice/crio-3aca8ca6d2fefb8303a31399852955c4178f4d0db9efda1c40031d5dd3bdda61 WatchSource:0}: Error finding container 3aca8ca6d2fefb8303a31399852955c4178f4d0db9efda1c40031d5dd3bdda61: Status 404 returned error can't find the container with id 3aca8ca6d2fefb8303a31399852955c4178f4d0db9efda1c40031d5dd3bdda61 Sep 30 08:05:33 crc kubenswrapper[4810]: I0930 08:05:33.125577 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"48ddf04a-273c-491e-a152-a67ef31eb463","Type":"ContainerDied","Data":"8f93a9758ca9a4c0a67fa556380c6b269c970283bdd0a66ab8e032d8a6f4b1bc"} Sep 30 08:05:33 crc kubenswrapper[4810]: I0930 08:05:33.125627 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f93a9758ca9a4c0a67fa556380c6b269c970283bdd0a66ab8e032d8a6f4b1bc" Sep 30 08:05:33 crc kubenswrapper[4810]: I0930 08:05:33.125720 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 08:05:33 crc kubenswrapper[4810]: I0930 08:05:33.164405 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3364826a-a3b8-43ec-b28a-4e616f064cf4","Type":"ContainerStarted","Data":"3aca8ca6d2fefb8303a31399852955c4178f4d0db9efda1c40031d5dd3bdda61"} Sep 30 08:05:34 crc kubenswrapper[4810]: I0930 08:05:34.321008 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3364826a-a3b8-43ec-b28a-4e616f064cf4","Type":"ContainerStarted","Data":"38182ccf81b3208de93e427eaeceb158e659ea886731e2fd44f6197e8ca40b2f"} Sep 30 08:05:34 crc kubenswrapper[4810]: I0930 08:05:34.356414 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.356387278 podStartE2EDuration="2.356387278s" podCreationTimestamp="2025-09-30 08:05:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:34.341753995 +0000 UTC m=+157.793953262" watchObservedRunningTime="2025-09-30 08:05:34.356387278 +0000 UTC m=+157.808586545" Sep 30 08:05:35 crc kubenswrapper[4810]: I0930 08:05:35.127248 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-6lcfr" Sep 30 08:05:35 crc kubenswrapper[4810]: I0930 08:05:35.355970 4810 generic.go:334] "Generic (PLEG): container finished" podID="3364826a-a3b8-43ec-b28a-4e616f064cf4" containerID="38182ccf81b3208de93e427eaeceb158e659ea886731e2fd44f6197e8ca40b2f" exitCode=0 Sep 30 08:05:35 crc kubenswrapper[4810]: I0930 08:05:35.356048 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3364826a-a3b8-43ec-b28a-4e616f064cf4","Type":"ContainerDied","Data":"38182ccf81b3208de93e427eaeceb158e659ea886731e2fd44f6197e8ca40b2f"} Sep 30 08:05:39 crc kubenswrapper[4810]: I0930 08:05:39.126038 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:39 crc kubenswrapper[4810]: I0930 08:05:39.141634 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:05:39 crc kubenswrapper[4810]: I0930 08:05:39.587238 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-47kbw" Sep 30 08:05:40 crc kubenswrapper[4810]: I0930 08:05:40.343389 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:05:40 crc kubenswrapper[4810]: I0930 08:05:40.353204 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/136360fc-f06a-4b28-bbe6-b8cefcac4fda-metrics-certs\") pod \"network-metrics-daemon-pnxwm\" (UID: \"136360fc-f06a-4b28-bbe6-b8cefcac4fda\") " pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:05:40 crc kubenswrapper[4810]: I0930 08:05:40.438326 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pnxwm" Sep 30 08:05:45 crc kubenswrapper[4810]: I0930 08:05:45.912439 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:05:45 crc kubenswrapper[4810]: I0930 08:05:45.913111 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:05:46 crc kubenswrapper[4810]: I0930 08:05:46.798363 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:05:50 crc kubenswrapper[4810]: I0930 08:05:50.162358 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 08:05:50 crc kubenswrapper[4810]: I0930 08:05:50.207717 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3364826a-a3b8-43ec-b28a-4e616f064cf4-kubelet-dir\") pod \"3364826a-a3b8-43ec-b28a-4e616f064cf4\" (UID: \"3364826a-a3b8-43ec-b28a-4e616f064cf4\") " Sep 30 08:05:50 crc kubenswrapper[4810]: I0930 08:05:50.207782 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3364826a-a3b8-43ec-b28a-4e616f064cf4-kube-api-access\") pod \"3364826a-a3b8-43ec-b28a-4e616f064cf4\" (UID: \"3364826a-a3b8-43ec-b28a-4e616f064cf4\") " Sep 30 08:05:50 crc kubenswrapper[4810]: I0930 08:05:50.207898 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3364826a-a3b8-43ec-b28a-4e616f064cf4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3364826a-a3b8-43ec-b28a-4e616f064cf4" (UID: "3364826a-a3b8-43ec-b28a-4e616f064cf4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:05:50 crc kubenswrapper[4810]: I0930 08:05:50.208330 4810 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3364826a-a3b8-43ec-b28a-4e616f064cf4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 30 08:05:50 crc kubenswrapper[4810]: I0930 08:05:50.230817 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3364826a-a3b8-43ec-b28a-4e616f064cf4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3364826a-a3b8-43ec-b28a-4e616f064cf4" (UID: "3364826a-a3b8-43ec-b28a-4e616f064cf4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:05:50 crc kubenswrapper[4810]: I0930 08:05:50.310036 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3364826a-a3b8-43ec-b28a-4e616f064cf4-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 08:05:50 crc kubenswrapper[4810]: I0930 08:05:50.492373 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3364826a-a3b8-43ec-b28a-4e616f064cf4","Type":"ContainerDied","Data":"3aca8ca6d2fefb8303a31399852955c4178f4d0db9efda1c40031d5dd3bdda61"} Sep 30 08:05:50 crc kubenswrapper[4810]: I0930 08:05:50.492440 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3aca8ca6d2fefb8303a31399852955c4178f4d0db9efda1c40031d5dd3bdda61" Sep 30 08:05:50 crc kubenswrapper[4810]: I0930 08:05:50.492471 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 08:05:56 crc kubenswrapper[4810]: E0930 08:05:56.539223 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Sep 30 08:05:56 crc kubenswrapper[4810]: E0930 08:05:56.540250 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wjl8c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-nqdvv_openshift-marketplace(8e6d85f7-29e2-41a8-b65e-02361a42f81d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 08:05:56 crc kubenswrapper[4810]: E0930 08:05:56.541760 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-nqdvv" podUID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" Sep 30 08:05:56 crc kubenswrapper[4810]: I0930 08:05:56.989519 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-pnxwm"] Sep 30 08:05:57 crc kubenswrapper[4810]: W0930 08:05:57.097155 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod136360fc_f06a_4b28_bbe6_b8cefcac4fda.slice/crio-1e2f002d3656371f7762a90c0853b7c03833ea7ecbd7fa186534a0c648a62b02 WatchSource:0}: Error finding container 1e2f002d3656371f7762a90c0853b7c03833ea7ecbd7fa186534a0c648a62b02: Status 404 returned error can't find the container with id 1e2f002d3656371f7762a90c0853b7c03833ea7ecbd7fa186534a0c648a62b02 Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.537015 4810 generic.go:334] "Generic (PLEG): container finished" podID="015f2a85-90c3-4bd9-b77a-63209e20cc74" containerID="5ddeecc1635379a06d57433c5a2a2770697b1e81c85508cd3e5d359778c00df0" exitCode=0 Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.537133 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vwd5" event={"ID":"015f2a85-90c3-4bd9-b77a-63209e20cc74","Type":"ContainerDied","Data":"5ddeecc1635379a06d57433c5a2a2770697b1e81c85508cd3e5d359778c00df0"} Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.544186 4810 generic.go:334] "Generic (PLEG): container finished" podID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" containerID="80e68ade2286c5751793707ae43ac92711e2eb74f4626ac20edfa555d40265a7" exitCode=0 Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.544362 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqwr" event={"ID":"fb1baa89-d430-4243-9b2d-2d2932c1abe4","Type":"ContainerDied","Data":"80e68ade2286c5751793707ae43ac92711e2eb74f4626ac20edfa555d40265a7"} Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.549202 4810 generic.go:334] "Generic (PLEG): container finished" podID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" containerID="cc723204f5e8427af71519452e9bd4239eba631def06c91d62bbcd413a5cb88c" exitCode=0 Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.549306 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrxfp" event={"ID":"76f9c24e-1aa7-4751-84c5-1a67afcbf75b","Type":"ContainerDied","Data":"cc723204f5e8427af71519452e9bd4239eba631def06c91d62bbcd413a5cb88c"} Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.553416 4810 generic.go:334] "Generic (PLEG): container finished" podID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" containerID="55939f1fb00cd55a38e93707977d076e5eb09b6b4c8d8aa364b3532f3f893643" exitCode=0 Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.553472 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8pf8" event={"ID":"51ccfc9f-81fc-4c77-95bb-fdb5e2239829","Type":"ContainerDied","Data":"55939f1fb00cd55a38e93707977d076e5eb09b6b4c8d8aa364b3532f3f893643"} Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.557448 4810 generic.go:334] "Generic (PLEG): container finished" podID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" containerID="7018509dad704fa1f12dfe394702b01380dd65a2edb40fddaab6e5723fd4e885" exitCode=0 Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.557522 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8x56x" event={"ID":"fee7bb01-669e-49de-b6fa-7cf3951fdfc2","Type":"ContainerDied","Data":"7018509dad704fa1f12dfe394702b01380dd65a2edb40fddaab6e5723fd4e885"} Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.572742 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" event={"ID":"136360fc-f06a-4b28-bbe6-b8cefcac4fda","Type":"ContainerStarted","Data":"1e2f002d3656371f7762a90c0853b7c03833ea7ecbd7fa186534a0c648a62b02"} Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.576475 4810 generic.go:334] "Generic (PLEG): container finished" podID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" containerID="10fee4f9b337f1cccf1ae20815cd3456a1f56632823574c5ebcae0b314426295" exitCode=0 Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.576615 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m27w2" event={"ID":"cb5387a0-ff9e-42bc-85b4-16e50a6f19de","Type":"ContainerDied","Data":"10fee4f9b337f1cccf1ae20815cd3456a1f56632823574c5ebcae0b314426295"} Sep 30 08:05:57 crc kubenswrapper[4810]: I0930 08:05:57.579045 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wprmn" event={"ID":"29a5cb45-7f5b-4700-b80c-3c60a0118d8c","Type":"ContainerStarted","Data":"a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3"} Sep 30 08:05:57 crc kubenswrapper[4810]: E0930 08:05:57.583747 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-nqdvv" podUID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" Sep 30 08:05:58 crc kubenswrapper[4810]: I0930 08:05:58.591037 4810 generic.go:334] "Generic (PLEG): container finished" podID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" containerID="a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3" exitCode=0 Sep 30 08:05:58 crc kubenswrapper[4810]: I0930 08:05:58.591140 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wprmn" event={"ID":"29a5cb45-7f5b-4700-b80c-3c60a0118d8c","Type":"ContainerDied","Data":"a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3"} Sep 30 08:05:58 crc kubenswrapper[4810]: I0930 08:05:58.594682 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" event={"ID":"136360fc-f06a-4b28-bbe6-b8cefcac4fda","Type":"ContainerStarted","Data":"6ba22f09e985fb1aa11c7732b1668fabc7f90eaa3bd08c8b51f1709d50a4e49b"} Sep 30 08:05:59 crc kubenswrapper[4810]: I0930 08:05:59.606361 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pnxwm" event={"ID":"136360fc-f06a-4b28-bbe6-b8cefcac4fda","Type":"ContainerStarted","Data":"62d59ceff819ed988f277a5dd2d923ada97c0e65117610e7207b06805875cf8a"} Sep 30 08:05:59 crc kubenswrapper[4810]: I0930 08:05:59.638800 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-pnxwm" podStartSLOduration=162.638761898 podStartE2EDuration="2m42.638761898s" podCreationTimestamp="2025-09-30 08:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:05:59.633360866 +0000 UTC m=+183.085560163" watchObservedRunningTime="2025-09-30 08:05:59.638761898 +0000 UTC m=+183.090961206" Sep 30 08:05:59 crc kubenswrapper[4810]: I0930 08:05:59.787693 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blbjc" Sep 30 08:06:00 crc kubenswrapper[4810]: I0930 08:06:00.618906 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8pf8" event={"ID":"51ccfc9f-81fc-4c77-95bb-fdb5e2239829","Type":"ContainerStarted","Data":"d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101"} Sep 30 08:06:00 crc kubenswrapper[4810]: I0930 08:06:00.649418 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t8pf8" podStartSLOduration=3.945643912 podStartE2EDuration="35.649383888s" podCreationTimestamp="2025-09-30 08:05:25 +0000 UTC" firstStartedPulling="2025-09-30 08:05:27.855606867 +0000 UTC m=+151.307806134" lastFinishedPulling="2025-09-30 08:05:59.559346803 +0000 UTC m=+183.011546110" observedRunningTime="2025-09-30 08:06:00.64733454 +0000 UTC m=+184.099533847" watchObservedRunningTime="2025-09-30 08:06:00.649383888 +0000 UTC m=+184.101583195" Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.637195 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqwr" event={"ID":"fb1baa89-d430-4243-9b2d-2d2932c1abe4","Type":"ContainerStarted","Data":"802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021"} Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.641729 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrxfp" event={"ID":"76f9c24e-1aa7-4751-84c5-1a67afcbf75b","Type":"ContainerStarted","Data":"370ee0d1e3f28b88662554ea2ff6a80cd0e07c635e86b6c9038ebba0aec32861"} Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.645226 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8x56x" event={"ID":"fee7bb01-669e-49de-b6fa-7cf3951fdfc2","Type":"ContainerStarted","Data":"ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb"} Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.648381 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m27w2" event={"ID":"cb5387a0-ff9e-42bc-85b4-16e50a6f19de","Type":"ContainerStarted","Data":"370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2"} Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.650875 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wprmn" event={"ID":"29a5cb45-7f5b-4700-b80c-3c60a0118d8c","Type":"ContainerStarted","Data":"64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421"} Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.653604 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vwd5" event={"ID":"015f2a85-90c3-4bd9-b77a-63209e20cc74","Type":"ContainerStarted","Data":"4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750"} Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.721234 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m27w2" podStartSLOduration=3.366180618 podStartE2EDuration="36.721210236s" podCreationTimestamp="2025-09-30 08:05:26 +0000 UTC" firstStartedPulling="2025-09-30 08:05:27.891856151 +0000 UTC m=+151.344055428" lastFinishedPulling="2025-09-30 08:06:01.246885769 +0000 UTC m=+184.699085046" observedRunningTime="2025-09-30 08:06:02.720421924 +0000 UTC m=+186.172621191" watchObservedRunningTime="2025-09-30 08:06:02.721210236 +0000 UTC m=+186.173409493" Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.723208 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xcqwr" podStartSLOduration=3.222370753 podStartE2EDuration="35.723200623s" podCreationTimestamp="2025-09-30 08:05:27 +0000 UTC" firstStartedPulling="2025-09-30 08:05:28.974414104 +0000 UTC m=+152.426613371" lastFinishedPulling="2025-09-30 08:06:01.475243934 +0000 UTC m=+184.927443241" observedRunningTime="2025-09-30 08:06:02.682335818 +0000 UTC m=+186.134535085" watchObservedRunningTime="2025-09-30 08:06:02.723200623 +0000 UTC m=+186.175399890" Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.743210 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lrxfp" podStartSLOduration=2.96365358 podStartE2EDuration="36.743188638s" podCreationTimestamp="2025-09-30 08:05:26 +0000 UTC" firstStartedPulling="2025-09-30 08:05:27.846225691 +0000 UTC m=+151.298424958" lastFinishedPulling="2025-09-30 08:06:01.625760729 +0000 UTC m=+185.077960016" observedRunningTime="2025-09-30 08:06:02.741587993 +0000 UTC m=+186.193787260" watchObservedRunningTime="2025-09-30 08:06:02.743188638 +0000 UTC m=+186.195387905" Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.802765 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8x56x" podStartSLOduration=4.403337787 podStartE2EDuration="34.802723591s" podCreationTimestamp="2025-09-30 08:05:28 +0000 UTC" firstStartedPulling="2025-09-30 08:05:31.09459754 +0000 UTC m=+154.546796807" lastFinishedPulling="2025-09-30 08:06:01.493983304 +0000 UTC m=+184.946182611" observedRunningTime="2025-09-30 08:06:02.772139576 +0000 UTC m=+186.224338853" watchObservedRunningTime="2025-09-30 08:06:02.802723591 +0000 UTC m=+186.254922848" Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.806012 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wprmn" podStartSLOduration=3.309953758 podStartE2EDuration="33.806005724s" podCreationTimestamp="2025-09-30 08:05:29 +0000 UTC" firstStartedPulling="2025-09-30 08:05:31.063500891 +0000 UTC m=+154.515700158" lastFinishedPulling="2025-09-30 08:06:01.559552867 +0000 UTC m=+185.011752124" observedRunningTime="2025-09-30 08:06:02.801704092 +0000 UTC m=+186.253903359" watchObservedRunningTime="2025-09-30 08:06:02.806005724 +0000 UTC m=+186.258204991" Sep 30 08:06:02 crc kubenswrapper[4810]: I0930 08:06:02.823809 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6vwd5" podStartSLOduration=2.983708237 podStartE2EDuration="37.823783476s" podCreationTimestamp="2025-09-30 08:05:25 +0000 UTC" firstStartedPulling="2025-09-30 08:05:26.844966036 +0000 UTC m=+150.297165303" lastFinishedPulling="2025-09-30 08:06:01.685041275 +0000 UTC m=+185.137240542" observedRunningTime="2025-09-30 08:06:02.820948316 +0000 UTC m=+186.273147583" watchObservedRunningTime="2025-09-30 08:06:02.823783476 +0000 UTC m=+186.275982743" Sep 30 08:06:05 crc kubenswrapper[4810]: I0930 08:06:05.357087 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.058648 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.060477 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.281727 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.281811 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.435440 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.435511 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.640542 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.641109 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.645388 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.691582 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.691930 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.747521 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.748548 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.748942 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:06:06 crc kubenswrapper[4810]: I0930 08:06:06.764137 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:06:07 crc kubenswrapper[4810]: I0930 08:06:07.756589 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:06:08 crc kubenswrapper[4810]: I0930 08:06:08.191492 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:06:08 crc kubenswrapper[4810]: I0930 08:06:08.191853 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:06:08 crc kubenswrapper[4810]: I0930 08:06:08.231959 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:06:08 crc kubenswrapper[4810]: I0930 08:06:08.801875 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:06:09 crc kubenswrapper[4810]: I0930 08:06:09.232433 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:06:09 crc kubenswrapper[4810]: I0930 08:06:09.232501 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:06:09 crc kubenswrapper[4810]: I0930 08:06:09.295460 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:06:09 crc kubenswrapper[4810]: I0930 08:06:09.594734 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:06:09 crc kubenswrapper[4810]: I0930 08:06:09.594950 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:06:09 crc kubenswrapper[4810]: I0930 08:06:09.635598 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:06:09 crc kubenswrapper[4810]: I0930 08:06:09.773009 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:06:09 crc kubenswrapper[4810]: I0930 08:06:09.795239 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:06:09 crc kubenswrapper[4810]: I0930 08:06:09.878163 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m27w2"] Sep 30 08:06:09 crc kubenswrapper[4810]: I0930 08:06:09.878496 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m27w2" podUID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" containerName="registry-server" containerID="cri-o://370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2" gracePeriod=2 Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.530501 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.648393 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwnn8\" (UniqueName: \"kubernetes.io/projected/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-kube-api-access-jwnn8\") pod \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.648498 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-catalog-content\") pod \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.648592 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-utilities\") pod \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\" (UID: \"cb5387a0-ff9e-42bc-85b4-16e50a6f19de\") " Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.650130 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-utilities" (OuterVolumeSpecName: "utilities") pod "cb5387a0-ff9e-42bc-85b4-16e50a6f19de" (UID: "cb5387a0-ff9e-42bc-85b4-16e50a6f19de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.671581 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-kube-api-access-jwnn8" (OuterVolumeSpecName: "kube-api-access-jwnn8") pod "cb5387a0-ff9e-42bc-85b4-16e50a6f19de" (UID: "cb5387a0-ff9e-42bc-85b4-16e50a6f19de"). InnerVolumeSpecName "kube-api-access-jwnn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.711741 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb5387a0-ff9e-42bc-85b4-16e50a6f19de" (UID: "cb5387a0-ff9e-42bc-85b4-16e50a6f19de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.725480 4810 generic.go:334] "Generic (PLEG): container finished" podID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" containerID="370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2" exitCode=0 Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.726537 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m27w2" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.727470 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m27w2" event={"ID":"cb5387a0-ff9e-42bc-85b4-16e50a6f19de","Type":"ContainerDied","Data":"370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2"} Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.727550 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m27w2" event={"ID":"cb5387a0-ff9e-42bc-85b4-16e50a6f19de","Type":"ContainerDied","Data":"1ffe784c60f5532a1d867aec8d4c207e0c71f2d8d98dcb20f10d71a07ce5bb42"} Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.727576 4810 scope.go:117] "RemoveContainer" containerID="370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.750950 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwnn8\" (UniqueName: \"kubernetes.io/projected/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-kube-api-access-jwnn8\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.750999 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.751022 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5387a0-ff9e-42bc-85b4-16e50a6f19de-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.783760 4810 scope.go:117] "RemoveContainer" containerID="10fee4f9b337f1cccf1ae20815cd3456a1f56632823574c5ebcae0b314426295" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.785007 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m27w2"] Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.789191 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m27w2"] Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.810240 4810 scope.go:117] "RemoveContainer" containerID="5e3e1081aad279c411b4eb6b9d912295a83fc4bbe0d753b8c6b85de0b7ba413f" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.825841 4810 scope.go:117] "RemoveContainer" containerID="370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2" Sep 30 08:06:10 crc kubenswrapper[4810]: E0930 08:06:10.826382 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2\": container with ID starting with 370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2 not found: ID does not exist" containerID="370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.826454 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2"} err="failed to get container status \"370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2\": rpc error: code = NotFound desc = could not find container \"370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2\": container with ID starting with 370c57da491a90ea61be535a4fe3173c04edcb883b76ce2e48ffa5cf577525a2 not found: ID does not exist" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.826524 4810 scope.go:117] "RemoveContainer" containerID="10fee4f9b337f1cccf1ae20815cd3456a1f56632823574c5ebcae0b314426295" Sep 30 08:06:10 crc kubenswrapper[4810]: E0930 08:06:10.826892 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10fee4f9b337f1cccf1ae20815cd3456a1f56632823574c5ebcae0b314426295\": container with ID starting with 10fee4f9b337f1cccf1ae20815cd3456a1f56632823574c5ebcae0b314426295 not found: ID does not exist" containerID="10fee4f9b337f1cccf1ae20815cd3456a1f56632823574c5ebcae0b314426295" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.826933 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10fee4f9b337f1cccf1ae20815cd3456a1f56632823574c5ebcae0b314426295"} err="failed to get container status \"10fee4f9b337f1cccf1ae20815cd3456a1f56632823574c5ebcae0b314426295\": rpc error: code = NotFound desc = could not find container \"10fee4f9b337f1cccf1ae20815cd3456a1f56632823574c5ebcae0b314426295\": container with ID starting with 10fee4f9b337f1cccf1ae20815cd3456a1f56632823574c5ebcae0b314426295 not found: ID does not exist" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.826959 4810 scope.go:117] "RemoveContainer" containerID="5e3e1081aad279c411b4eb6b9d912295a83fc4bbe0d753b8c6b85de0b7ba413f" Sep 30 08:06:10 crc kubenswrapper[4810]: E0930 08:06:10.827613 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e3e1081aad279c411b4eb6b9d912295a83fc4bbe0d753b8c6b85de0b7ba413f\": container with ID starting with 5e3e1081aad279c411b4eb6b9d912295a83fc4bbe0d753b8c6b85de0b7ba413f not found: ID does not exist" containerID="5e3e1081aad279c411b4eb6b9d912295a83fc4bbe0d753b8c6b85de0b7ba413f" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.827650 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e3e1081aad279c411b4eb6b9d912295a83fc4bbe0d753b8c6b85de0b7ba413f"} err="failed to get container status \"5e3e1081aad279c411b4eb6b9d912295a83fc4bbe0d753b8c6b85de0b7ba413f\": rpc error: code = NotFound desc = could not find container \"5e3e1081aad279c411b4eb6b9d912295a83fc4bbe0d753b8c6b85de0b7ba413f\": container with ID starting with 5e3e1081aad279c411b4eb6b9d912295a83fc4bbe0d753b8c6b85de0b7ba413f not found: ID does not exist" Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.882873 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lrxfp"] Sep 30 08:06:10 crc kubenswrapper[4810]: I0930 08:06:10.883391 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lrxfp" podUID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" containerName="registry-server" containerID="cri-o://370ee0d1e3f28b88662554ea2ff6a80cd0e07c635e86b6c9038ebba0aec32861" gracePeriod=2 Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.318884 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" path="/var/lib/kubelet/pods/cb5387a0-ff9e-42bc-85b4-16e50a6f19de/volumes" Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.738620 4810 generic.go:334] "Generic (PLEG): container finished" podID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" containerID="370ee0d1e3f28b88662554ea2ff6a80cd0e07c635e86b6c9038ebba0aec32861" exitCode=0 Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.738740 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrxfp" event={"ID":"76f9c24e-1aa7-4751-84c5-1a67afcbf75b","Type":"ContainerDied","Data":"370ee0d1e3f28b88662554ea2ff6a80cd0e07c635e86b6c9038ebba0aec32861"} Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.780759 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.882529 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls599\" (UniqueName: \"kubernetes.io/projected/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-kube-api-access-ls599\") pod \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.882630 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-catalog-content\") pod \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.890549 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-kube-api-access-ls599" (OuterVolumeSpecName: "kube-api-access-ls599") pod "76f9c24e-1aa7-4751-84c5-1a67afcbf75b" (UID: "76f9c24e-1aa7-4751-84c5-1a67afcbf75b"). InnerVolumeSpecName "kube-api-access-ls599". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.894451 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-utilities\") pod \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\" (UID: \"76f9c24e-1aa7-4751-84c5-1a67afcbf75b\") " Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.895057 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls599\" (UniqueName: \"kubernetes.io/projected/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-kube-api-access-ls599\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.895165 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-utilities" (OuterVolumeSpecName: "utilities") pod "76f9c24e-1aa7-4751-84c5-1a67afcbf75b" (UID: "76f9c24e-1aa7-4751-84c5-1a67afcbf75b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.953434 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76f9c24e-1aa7-4751-84c5-1a67afcbf75b" (UID: "76f9c24e-1aa7-4751-84c5-1a67afcbf75b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.995936 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:11 crc kubenswrapper[4810]: I0930 08:06:11.995988 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76f9c24e-1aa7-4751-84c5-1a67afcbf75b-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:12 crc kubenswrapper[4810]: I0930 08:06:12.749378 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqdvv" event={"ID":"8e6d85f7-29e2-41a8-b65e-02361a42f81d","Type":"ContainerStarted","Data":"99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9"} Sep 30 08:06:12 crc kubenswrapper[4810]: I0930 08:06:12.753345 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrxfp" event={"ID":"76f9c24e-1aa7-4751-84c5-1a67afcbf75b","Type":"ContainerDied","Data":"22c7ba6e8b7bb503cb2b84fdfe598ddcea2b64ff0e8b96527893747e9ac3987d"} Sep 30 08:06:12 crc kubenswrapper[4810]: I0930 08:06:12.753392 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lrxfp" Sep 30 08:06:12 crc kubenswrapper[4810]: I0930 08:06:12.753395 4810 scope.go:117] "RemoveContainer" containerID="370ee0d1e3f28b88662554ea2ff6a80cd0e07c635e86b6c9038ebba0aec32861" Sep 30 08:06:12 crc kubenswrapper[4810]: I0930 08:06:12.780916 4810 scope.go:117] "RemoveContainer" containerID="cc723204f5e8427af71519452e9bd4239eba631def06c91d62bbcd413a5cb88c" Sep 30 08:06:12 crc kubenswrapper[4810]: I0930 08:06:12.786438 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lrxfp"] Sep 30 08:06:12 crc kubenswrapper[4810]: I0930 08:06:12.790989 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lrxfp"] Sep 30 08:06:12 crc kubenswrapper[4810]: I0930 08:06:12.850888 4810 scope.go:117] "RemoveContainer" containerID="359e775036370e0921095f82a9f45a3b857f47b99e122aadb3991e995d4f48b2" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.285084 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wprmn"] Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.287167 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wprmn" podUID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" containerName="registry-server" containerID="cri-o://64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421" gracePeriod=2 Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.318667 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" path="/var/lib/kubelet/pods/76f9c24e-1aa7-4751-84c5-1a67afcbf75b/volumes" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.726357 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.768190 4810 generic.go:334] "Generic (PLEG): container finished" podID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" containerID="64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421" exitCode=0 Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.768240 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wprmn" event={"ID":"29a5cb45-7f5b-4700-b80c-3c60a0118d8c","Type":"ContainerDied","Data":"64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421"} Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.768295 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wprmn" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.768347 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wprmn" event={"ID":"29a5cb45-7f5b-4700-b80c-3c60a0118d8c","Type":"ContainerDied","Data":"a4d70b622f1a0f2add55fe204b1bb85c88b7188795e1d1ec0c4d668fb813cab5"} Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.768376 4810 scope.go:117] "RemoveContainer" containerID="64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.770774 4810 generic.go:334] "Generic (PLEG): container finished" podID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" containerID="99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9" exitCode=0 Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.770819 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqdvv" event={"ID":"8e6d85f7-29e2-41a8-b65e-02361a42f81d","Type":"ContainerDied","Data":"99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9"} Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.789815 4810 scope.go:117] "RemoveContainer" containerID="a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.804340 4810 scope.go:117] "RemoveContainer" containerID="f0f06aeb11843e40a0f90e83f436a5a9c736e44ae35f5cb92ee45e57b41b3fca" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.823543 4810 scope.go:117] "RemoveContainer" containerID="64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421" Sep 30 08:06:13 crc kubenswrapper[4810]: E0930 08:06:13.824349 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421\": container with ID starting with 64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421 not found: ID does not exist" containerID="64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.824399 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421"} err="failed to get container status \"64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421\": rpc error: code = NotFound desc = could not find container \"64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421\": container with ID starting with 64c9ab0c6669e01022f773254617ccac396f2799194ce1c152b839f606f7a421 not found: ID does not exist" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.824429 4810 scope.go:117] "RemoveContainer" containerID="a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3" Sep 30 08:06:13 crc kubenswrapper[4810]: E0930 08:06:13.824773 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3\": container with ID starting with a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3 not found: ID does not exist" containerID="a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.824799 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3"} err="failed to get container status \"a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3\": rpc error: code = NotFound desc = could not find container \"a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3\": container with ID starting with a8dfed72894fd20036e0bdabaed3d89809b62988934f7b0bfa95ca71580fc5a3 not found: ID does not exist" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.824812 4810 scope.go:117] "RemoveContainer" containerID="f0f06aeb11843e40a0f90e83f436a5a9c736e44ae35f5cb92ee45e57b41b3fca" Sep 30 08:06:13 crc kubenswrapper[4810]: E0930 08:06:13.825166 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0f06aeb11843e40a0f90e83f436a5a9c736e44ae35f5cb92ee45e57b41b3fca\": container with ID starting with f0f06aeb11843e40a0f90e83f436a5a9c736e44ae35f5cb92ee45e57b41b3fca not found: ID does not exist" containerID="f0f06aeb11843e40a0f90e83f436a5a9c736e44ae35f5cb92ee45e57b41b3fca" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.825192 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0f06aeb11843e40a0f90e83f436a5a9c736e44ae35f5cb92ee45e57b41b3fca"} err="failed to get container status \"f0f06aeb11843e40a0f90e83f436a5a9c736e44ae35f5cb92ee45e57b41b3fca\": rpc error: code = NotFound desc = could not find container \"f0f06aeb11843e40a0f90e83f436a5a9c736e44ae35f5cb92ee45e57b41b3fca\": container with ID starting with f0f06aeb11843e40a0f90e83f436a5a9c736e44ae35f5cb92ee45e57b41b3fca not found: ID does not exist" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.925188 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-utilities\") pod \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.926484 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-utilities" (OuterVolumeSpecName: "utilities") pod "29a5cb45-7f5b-4700-b80c-3c60a0118d8c" (UID: "29a5cb45-7f5b-4700-b80c-3c60a0118d8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.926672 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-catalog-content\") pod \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.941476 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brpsd\" (UniqueName: \"kubernetes.io/projected/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-kube-api-access-brpsd\") pod \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\" (UID: \"29a5cb45-7f5b-4700-b80c-3c60a0118d8c\") " Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.941967 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:13 crc kubenswrapper[4810]: I0930 08:06:13.949433 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-kube-api-access-brpsd" (OuterVolumeSpecName: "kube-api-access-brpsd") pod "29a5cb45-7f5b-4700-b80c-3c60a0118d8c" (UID: "29a5cb45-7f5b-4700-b80c-3c60a0118d8c"). InnerVolumeSpecName "kube-api-access-brpsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:06:14 crc kubenswrapper[4810]: I0930 08:06:14.015816 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29a5cb45-7f5b-4700-b80c-3c60a0118d8c" (UID: "29a5cb45-7f5b-4700-b80c-3c60a0118d8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:14 crc kubenswrapper[4810]: I0930 08:06:14.042835 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:14 crc kubenswrapper[4810]: I0930 08:06:14.042886 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brpsd\" (UniqueName: \"kubernetes.io/projected/29a5cb45-7f5b-4700-b80c-3c60a0118d8c-kube-api-access-brpsd\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:14 crc kubenswrapper[4810]: I0930 08:06:14.114680 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wprmn"] Sep 30 08:06:14 crc kubenswrapper[4810]: I0930 08:06:14.118550 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wprmn"] Sep 30 08:06:14 crc kubenswrapper[4810]: I0930 08:06:14.790536 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqdvv" event={"ID":"8e6d85f7-29e2-41a8-b65e-02361a42f81d","Type":"ContainerStarted","Data":"d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c"} Sep 30 08:06:14 crc kubenswrapper[4810]: I0930 08:06:14.812581 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nqdvv" podStartSLOduration=2.618411638 podStartE2EDuration="46.812563483s" podCreationTimestamp="2025-09-30 08:05:28 +0000 UTC" firstStartedPulling="2025-09-30 08:05:30.003481645 +0000 UTC m=+153.455680912" lastFinishedPulling="2025-09-30 08:06:14.19763349 +0000 UTC m=+197.649832757" observedRunningTime="2025-09-30 08:06:14.809844133 +0000 UTC m=+198.262043400" watchObservedRunningTime="2025-09-30 08:06:14.812563483 +0000 UTC m=+198.264762750" Sep 30 08:06:15 crc kubenswrapper[4810]: I0930 08:06:15.315676 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" path="/var/lib/kubelet/pods/29a5cb45-7f5b-4700-b80c-3c60a0118d8c/volumes" Sep 30 08:06:15 crc kubenswrapper[4810]: I0930 08:06:15.911419 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:06:15 crc kubenswrapper[4810]: I0930 08:06:15.911496 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:06:18 crc kubenswrapper[4810]: I0930 08:06:18.582795 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:06:18 crc kubenswrapper[4810]: I0930 08:06:18.583066 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:06:18 crc kubenswrapper[4810]: I0930 08:06:18.619475 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:06:28 crc kubenswrapper[4810]: I0930 08:06:28.622402 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.079882 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqdvv"] Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.080568 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nqdvv" podUID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" containerName="registry-server" containerID="cri-o://d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c" gracePeriod=2 Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.404724 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.569615 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-utilities\") pod \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.569697 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjl8c\" (UniqueName: \"kubernetes.io/projected/8e6d85f7-29e2-41a8-b65e-02361a42f81d-kube-api-access-wjl8c\") pod \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.569787 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-catalog-content\") pod \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\" (UID: \"8e6d85f7-29e2-41a8-b65e-02361a42f81d\") " Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.570438 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-utilities" (OuterVolumeSpecName: "utilities") pod "8e6d85f7-29e2-41a8-b65e-02361a42f81d" (UID: "8e6d85f7-29e2-41a8-b65e-02361a42f81d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.575883 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e6d85f7-29e2-41a8-b65e-02361a42f81d-kube-api-access-wjl8c" (OuterVolumeSpecName: "kube-api-access-wjl8c") pod "8e6d85f7-29e2-41a8-b65e-02361a42f81d" (UID: "8e6d85f7-29e2-41a8-b65e-02361a42f81d"). InnerVolumeSpecName "kube-api-access-wjl8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.587464 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e6d85f7-29e2-41a8-b65e-02361a42f81d" (UID: "8e6d85f7-29e2-41a8-b65e-02361a42f81d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.671213 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.671277 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e6d85f7-29e2-41a8-b65e-02361a42f81d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.671291 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjl8c\" (UniqueName: \"kubernetes.io/projected/8e6d85f7-29e2-41a8-b65e-02361a42f81d-kube-api-access-wjl8c\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.918849 4810 generic.go:334] "Generic (PLEG): container finished" podID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" containerID="d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c" exitCode=0 Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.918895 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqdvv" event={"ID":"8e6d85f7-29e2-41a8-b65e-02361a42f81d","Type":"ContainerDied","Data":"d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c"} Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.918925 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqdvv" event={"ID":"8e6d85f7-29e2-41a8-b65e-02361a42f81d","Type":"ContainerDied","Data":"b47398b5e4aedb7b1fe6f78fded5bd0ae58552e56bea54613269bbe023ff68d5"} Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.918944 4810 scope.go:117] "RemoveContainer" containerID="d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.918942 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqdvv" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.940193 4810 scope.go:117] "RemoveContainer" containerID="99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.947148 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqdvv"] Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.950032 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqdvv"] Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.957597 4810 scope.go:117] "RemoveContainer" containerID="ec0ac67a412f96254510f299f8521ef3a4d88ce06cea4933c7db87b5d7400ec8" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.989693 4810 scope.go:117] "RemoveContainer" containerID="d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c" Sep 30 08:06:31 crc kubenswrapper[4810]: E0930 08:06:31.990167 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c\": container with ID starting with d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c not found: ID does not exist" containerID="d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.990225 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c"} err="failed to get container status \"d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c\": rpc error: code = NotFound desc = could not find container \"d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c\": container with ID starting with d24887417f953d9cae075ae4bd470e912e86a7e714a2a41035e15bda7213f04c not found: ID does not exist" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.990303 4810 scope.go:117] "RemoveContainer" containerID="99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9" Sep 30 08:06:31 crc kubenswrapper[4810]: E0930 08:06:31.990610 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9\": container with ID starting with 99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9 not found: ID does not exist" containerID="99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.990666 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9"} err="failed to get container status \"99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9\": rpc error: code = NotFound desc = could not find container \"99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9\": container with ID starting with 99e0c876cfe37b56f18e37220a247291eac22e619d0dccda9f4e8c2c346d26f9 not found: ID does not exist" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.990702 4810 scope.go:117] "RemoveContainer" containerID="ec0ac67a412f96254510f299f8521ef3a4d88ce06cea4933c7db87b5d7400ec8" Sep 30 08:06:31 crc kubenswrapper[4810]: E0930 08:06:31.991112 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec0ac67a412f96254510f299f8521ef3a4d88ce06cea4933c7db87b5d7400ec8\": container with ID starting with ec0ac67a412f96254510f299f8521ef3a4d88ce06cea4933c7db87b5d7400ec8 not found: ID does not exist" containerID="ec0ac67a412f96254510f299f8521ef3a4d88ce06cea4933c7db87b5d7400ec8" Sep 30 08:06:31 crc kubenswrapper[4810]: I0930 08:06:31.991147 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec0ac67a412f96254510f299f8521ef3a4d88ce06cea4933c7db87b5d7400ec8"} err="failed to get container status \"ec0ac67a412f96254510f299f8521ef3a4d88ce06cea4933c7db87b5d7400ec8\": rpc error: code = NotFound desc = could not find container \"ec0ac67a412f96254510f299f8521ef3a4d88ce06cea4933c7db87b5d7400ec8\": container with ID starting with ec0ac67a412f96254510f299f8521ef3a4d88ce06cea4933c7db87b5d7400ec8 not found: ID does not exist" Sep 30 08:06:33 crc kubenswrapper[4810]: I0930 08:06:33.323525 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" path="/var/lib/kubelet/pods/8e6d85f7-29e2-41a8-b65e-02361a42f81d/volumes" Sep 30 08:06:45 crc kubenswrapper[4810]: I0930 08:06:45.911967 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:06:45 crc kubenswrapper[4810]: I0930 08:06:45.912667 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:06:45 crc kubenswrapper[4810]: I0930 08:06:45.912740 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:06:45 crc kubenswrapper[4810]: I0930 08:06:45.913587 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:06:45 crc kubenswrapper[4810]: I0930 08:06:45.913687 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4" gracePeriod=600 Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.009562 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6vwd5"] Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.010906 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6vwd5" podUID="015f2a85-90c3-4bd9-b77a-63209e20cc74" containerName="registry-server" containerID="cri-o://4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750" gracePeriod=30 Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.021399 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4" exitCode=0 Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.021470 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4"} Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.021723 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"33485b79f13ae3509d05af432da899150f9b5186a42141758f7a50e6033f5705"} Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.030459 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t8pf8"] Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.031159 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t8pf8" podUID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" containerName="registry-server" containerID="cri-o://d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101" gracePeriod=30 Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.048147 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tmk5l"] Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.048428 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" podUID="e049bda6-e38f-4357-926e-d8b767cf5ed9" containerName="marketplace-operator" containerID="cri-o://f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec" gracePeriod=30 Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.056082 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqwr"] Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.056471 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xcqwr" podUID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" containerName="registry-server" containerID="cri-o://802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021" gracePeriod=30 Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.059866 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8x56x"] Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.060170 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8x56x" podUID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" containerName="registry-server" containerID="cri-o://ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb" gracePeriod=30 Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.063417 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-txqd2"] Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.063832 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" containerName="extract-utilities" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.063858 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" containerName="extract-utilities" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.063876 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" containerName="extract-utilities" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.063885 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" containerName="extract-utilities" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.063898 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.063907 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.063921 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" containerName="extract-content" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.063931 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" containerName="extract-content" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.063948 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" containerName="extract-utilities" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.063959 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" containerName="extract-utilities" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.063972 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" containerName="extract-utilities" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.063982 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" containerName="extract-utilities" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.063991 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3364826a-a3b8-43ec-b28a-4e616f064cf4" containerName="pruner" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.063999 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3364826a-a3b8-43ec-b28a-4e616f064cf4" containerName="pruner" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.064011 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064020 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.064031 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ddf04a-273c-491e-a152-a67ef31eb463" containerName="pruner" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064040 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ddf04a-273c-491e-a152-a67ef31eb463" containerName="pruner" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.064049 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" containerName="extract-content" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064057 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" containerName="extract-content" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.064071 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064080 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.064092 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064100 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.064114 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" containerName="extract-content" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064123 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" containerName="extract-content" Sep 30 08:06:47 crc kubenswrapper[4810]: E0930 08:06:47.064135 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" containerName="extract-content" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064144 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" containerName="extract-content" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064317 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f9c24e-1aa7-4751-84c5-1a67afcbf75b" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064351 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a5cb45-7f5b-4700-b80c-3c60a0118d8c" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064366 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="48ddf04a-273c-491e-a152-a67ef31eb463" containerName="pruner" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064376 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb5387a0-ff9e-42bc-85b4-16e50a6f19de" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064387 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3364826a-a3b8-43ec-b28a-4e616f064cf4" containerName="pruner" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064399 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e6d85f7-29e2-41a8-b65e-02361a42f81d" containerName="registry-server" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.064993 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.068397 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-txqd2"] Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.086715 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-txqd2\" (UID: \"7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba\") " pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.086857 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktvp2\" (UniqueName: \"kubernetes.io/projected/7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba-kube-api-access-ktvp2\") pod \"marketplace-operator-79b997595-txqd2\" (UID: \"7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba\") " pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.086894 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-txqd2\" (UID: \"7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba\") " pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.188607 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktvp2\" (UniqueName: \"kubernetes.io/projected/7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba-kube-api-access-ktvp2\") pod \"marketplace-operator-79b997595-txqd2\" (UID: \"7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba\") " pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.188684 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-txqd2\" (UID: \"7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba\") " pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.188801 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-txqd2\" (UID: \"7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba\") " pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.191528 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-txqd2\" (UID: \"7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba\") " pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.197083 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-txqd2\" (UID: \"7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba\") " pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.213957 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktvp2\" (UniqueName: \"kubernetes.io/projected/7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba-kube-api-access-ktvp2\") pod \"marketplace-operator-79b997595-txqd2\" (UID: \"7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba\") " pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.384598 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.507964 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.528635 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.535735 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.561207 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.601564 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-catalog-content\") pod \"015f2a85-90c3-4bd9-b77a-63209e20cc74\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.601651 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ts6qk\" (UniqueName: \"kubernetes.io/projected/015f2a85-90c3-4bd9-b77a-63209e20cc74-kube-api-access-ts6qk\") pod \"015f2a85-90c3-4bd9-b77a-63209e20cc74\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.601673 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-utilities\") pod \"015f2a85-90c3-4bd9-b77a-63209e20cc74\" (UID: \"015f2a85-90c3-4bd9-b77a-63209e20cc74\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.602819 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-utilities" (OuterVolumeSpecName: "utilities") pod "015f2a85-90c3-4bd9-b77a-63209e20cc74" (UID: "015f2a85-90c3-4bd9-b77a-63209e20cc74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.607941 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/015f2a85-90c3-4bd9-b77a-63209e20cc74-kube-api-access-ts6qk" (OuterVolumeSpecName: "kube-api-access-ts6qk") pod "015f2a85-90c3-4bd9-b77a-63209e20cc74" (UID: "015f2a85-90c3-4bd9-b77a-63209e20cc74"). InnerVolumeSpecName "kube-api-access-ts6qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.653458 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "015f2a85-90c3-4bd9-b77a-63209e20cc74" (UID: "015f2a85-90c3-4bd9-b77a-63209e20cc74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.703074 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-operator-metrics\") pod \"e049bda6-e38f-4357-926e-d8b767cf5ed9\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.703151 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-utilities\") pod \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.703237 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-catalog-content\") pod \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.703289 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m6bg\" (UniqueName: \"kubernetes.io/projected/fb1baa89-d430-4243-9b2d-2d2932c1abe4-kube-api-access-4m6bg\") pod \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.703534 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-utilities\") pod \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\" (UID: \"fb1baa89-d430-4243-9b2d-2d2932c1abe4\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.703568 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-trusted-ca\") pod \"e049bda6-e38f-4357-926e-d8b767cf5ed9\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.703588 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5892\" (UniqueName: \"kubernetes.io/projected/e049bda6-e38f-4357-926e-d8b767cf5ed9-kube-api-access-m5892\") pod \"e049bda6-e38f-4357-926e-d8b767cf5ed9\" (UID: \"e049bda6-e38f-4357-926e-d8b767cf5ed9\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.703627 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b49p\" (UniqueName: \"kubernetes.io/projected/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-kube-api-access-2b49p\") pod \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.703648 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-catalog-content\") pod \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\" (UID: \"fee7bb01-669e-49de-b6fa-7cf3951fdfc2\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.704041 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.704056 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ts6qk\" (UniqueName: \"kubernetes.io/projected/015f2a85-90c3-4bd9-b77a-63209e20cc74-kube-api-access-ts6qk\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.704067 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015f2a85-90c3-4bd9-b77a-63209e20cc74-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.704750 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-utilities" (OuterVolumeSpecName: "utilities") pod "fee7bb01-669e-49de-b6fa-7cf3951fdfc2" (UID: "fee7bb01-669e-49de-b6fa-7cf3951fdfc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.706065 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "e049bda6-e38f-4357-926e-d8b767cf5ed9" (UID: "e049bda6-e38f-4357-926e-d8b767cf5ed9"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.707186 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb1baa89-d430-4243-9b2d-2d2932c1abe4-kube-api-access-4m6bg" (OuterVolumeSpecName: "kube-api-access-4m6bg") pod "fb1baa89-d430-4243-9b2d-2d2932c1abe4" (UID: "fb1baa89-d430-4243-9b2d-2d2932c1abe4"). InnerVolumeSpecName "kube-api-access-4m6bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.708133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e049bda6-e38f-4357-926e-d8b767cf5ed9-kube-api-access-m5892" (OuterVolumeSpecName: "kube-api-access-m5892") pod "e049bda6-e38f-4357-926e-d8b767cf5ed9" (UID: "e049bda6-e38f-4357-926e-d8b767cf5ed9"). InnerVolumeSpecName "kube-api-access-m5892". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.708726 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "e049bda6-e38f-4357-926e-d8b767cf5ed9" (UID: "e049bda6-e38f-4357-926e-d8b767cf5ed9"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.708925 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-kube-api-access-2b49p" (OuterVolumeSpecName: "kube-api-access-2b49p") pod "fee7bb01-669e-49de-b6fa-7cf3951fdfc2" (UID: "fee7bb01-669e-49de-b6fa-7cf3951fdfc2"). InnerVolumeSpecName "kube-api-access-2b49p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.710285 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-utilities" (OuterVolumeSpecName: "utilities") pod "fb1baa89-d430-4243-9b2d-2d2932c1abe4" (UID: "fb1baa89-d430-4243-9b2d-2d2932c1abe4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.724710 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.729324 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb1baa89-d430-4243-9b2d-2d2932c1abe4" (UID: "fb1baa89-d430-4243-9b2d-2d2932c1abe4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.799387 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fee7bb01-669e-49de-b6fa-7cf3951fdfc2" (UID: "fee7bb01-669e-49de-b6fa-7cf3951fdfc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.804858 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b49p\" (UniqueName: \"kubernetes.io/projected/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-kube-api-access-2b49p\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.804902 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.804915 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.804940 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fee7bb01-669e-49de-b6fa-7cf3951fdfc2-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.804950 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.805087 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m6bg\" (UniqueName: \"kubernetes.io/projected/fb1baa89-d430-4243-9b2d-2d2932c1abe4-kube-api-access-4m6bg\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.805099 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1baa89-d430-4243-9b2d-2d2932c1abe4-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.805111 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e049bda6-e38f-4357-926e-d8b767cf5ed9-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.805123 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5892\" (UniqueName: \"kubernetes.io/projected/e049bda6-e38f-4357-926e-d8b767cf5ed9-kube-api-access-m5892\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.833908 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-txqd2"] Sep 30 08:06:47 crc kubenswrapper[4810]: W0930 08:06:47.842134 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7aace4e5_3195_4ec0_b0cf_c44d85c1f5ba.slice/crio-1df00a1c5516b20f3f04db75d5e9701d08928267b4e2f5c3344bfa0c837542f6 WatchSource:0}: Error finding container 1df00a1c5516b20f3f04db75d5e9701d08928267b4e2f5c3344bfa0c837542f6: Status 404 returned error can't find the container with id 1df00a1c5516b20f3f04db75d5e9701d08928267b4e2f5c3344bfa0c837542f6 Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.905782 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-catalog-content\") pod \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.905819 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-utilities\") pod \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.905855 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxhpx\" (UniqueName: \"kubernetes.io/projected/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-kube-api-access-rxhpx\") pod \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\" (UID: \"51ccfc9f-81fc-4c77-95bb-fdb5e2239829\") " Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.906566 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-utilities" (OuterVolumeSpecName: "utilities") pod "51ccfc9f-81fc-4c77-95bb-fdb5e2239829" (UID: "51ccfc9f-81fc-4c77-95bb-fdb5e2239829"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.910748 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-kube-api-access-rxhpx" (OuterVolumeSpecName: "kube-api-access-rxhpx") pod "51ccfc9f-81fc-4c77-95bb-fdb5e2239829" (UID: "51ccfc9f-81fc-4c77-95bb-fdb5e2239829"). InnerVolumeSpecName "kube-api-access-rxhpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:06:47 crc kubenswrapper[4810]: I0930 08:06:47.965492 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51ccfc9f-81fc-4c77-95bb-fdb5e2239829" (UID: "51ccfc9f-81fc-4c77-95bb-fdb5e2239829"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.007855 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.007892 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.007907 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxhpx\" (UniqueName: \"kubernetes.io/projected/51ccfc9f-81fc-4c77-95bb-fdb5e2239829-kube-api-access-rxhpx\") on node \"crc\" DevicePath \"\"" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.027693 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" event={"ID":"7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba","Type":"ContainerStarted","Data":"e15f96b8b0f1b230c4e31f299100baa8a3e93d0cc48336104375be6b98779571"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.027742 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" event={"ID":"7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba","Type":"ContainerStarted","Data":"1df00a1c5516b20f3f04db75d5e9701d08928267b4e2f5c3344bfa0c837542f6"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.028151 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.029028 4810 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-txqd2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.54:8080/healthz\": dial tcp 10.217.0.54:8080: connect: connection refused" start-of-body= Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.029066 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" podUID="7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.54:8080/healthz\": dial tcp 10.217.0.54:8080: connect: connection refused" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.030474 4810 generic.go:334] "Generic (PLEG): container finished" podID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" containerID="ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb" exitCode=0 Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.030516 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8x56x" event={"ID":"fee7bb01-669e-49de-b6fa-7cf3951fdfc2","Type":"ContainerDied","Data":"ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.030535 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8x56x" event={"ID":"fee7bb01-669e-49de-b6fa-7cf3951fdfc2","Type":"ContainerDied","Data":"3e194d210b9ab7b15870ed46eab69bd4d7f124cb4e0d52d13d15068d56f58741"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.030552 4810 scope.go:117] "RemoveContainer" containerID="ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.030555 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8x56x" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.032532 4810 generic.go:334] "Generic (PLEG): container finished" podID="e049bda6-e38f-4357-926e-d8b767cf5ed9" containerID="f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec" exitCode=0 Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.032590 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.032600 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" event={"ID":"e049bda6-e38f-4357-926e-d8b767cf5ed9","Type":"ContainerDied","Data":"f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.032640 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tmk5l" event={"ID":"e049bda6-e38f-4357-926e-d8b767cf5ed9","Type":"ContainerDied","Data":"620392ec040254ddb29445b086753d82ccea05fa80f3226fa5c798e8531dfe29"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.035867 4810 generic.go:334] "Generic (PLEG): container finished" podID="015f2a85-90c3-4bd9-b77a-63209e20cc74" containerID="4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750" exitCode=0 Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.035921 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vwd5" event={"ID":"015f2a85-90c3-4bd9-b77a-63209e20cc74","Type":"ContainerDied","Data":"4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.035942 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vwd5" event={"ID":"015f2a85-90c3-4bd9-b77a-63209e20cc74","Type":"ContainerDied","Data":"45214645d40e376161531e77af5d98ccfaa2c3832f6a836c2f6c4e0e047df397"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.036020 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vwd5" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.043426 4810 generic.go:334] "Generic (PLEG): container finished" podID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" containerID="802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021" exitCode=0 Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.043553 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqwr" event={"ID":"fb1baa89-d430-4243-9b2d-2d2932c1abe4","Type":"ContainerDied","Data":"802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.043591 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqwr" event={"ID":"fb1baa89-d430-4243-9b2d-2d2932c1abe4","Type":"ContainerDied","Data":"cd78dd2ea82a1fb37445b03443d83e09fda67ba2c2e4eb926f431c84fb4418f4"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.043551 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xcqwr" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.046766 4810 generic.go:334] "Generic (PLEG): container finished" podID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" containerID="d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101" exitCode=0 Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.046812 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8pf8" event={"ID":"51ccfc9f-81fc-4c77-95bb-fdb5e2239829","Type":"ContainerDied","Data":"d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.046835 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t8pf8" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.046840 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t8pf8" event={"ID":"51ccfc9f-81fc-4c77-95bb-fdb5e2239829","Type":"ContainerDied","Data":"f26133c9396f16dbf2dc05d741d26ed1819de76354146e5981d587c965a0e271"} Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.057728 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" podStartSLOduration=1.057701065 podStartE2EDuration="1.057701065s" podCreationTimestamp="2025-09-30 08:06:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:06:48.04995119 +0000 UTC m=+231.502150467" watchObservedRunningTime="2025-09-30 08:06:48.057701065 +0000 UTC m=+231.509900342" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.076899 4810 scope.go:117] "RemoveContainer" containerID="7018509dad704fa1f12dfe394702b01380dd65a2edb40fddaab6e5723fd4e885" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.092755 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6vwd5"] Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.098751 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6vwd5"] Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.112012 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tmk5l"] Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.116895 4810 scope.go:117] "RemoveContainer" containerID="11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.128216 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tmk5l"] Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.144416 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t8pf8"] Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.158938 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t8pf8"] Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.159471 4810 scope.go:117] "RemoveContainer" containerID="ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.161237 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb\": container with ID starting with ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb not found: ID does not exist" containerID="ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.161318 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb"} err="failed to get container status \"ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb\": rpc error: code = NotFound desc = could not find container \"ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb\": container with ID starting with ccb7f133c3e84f35fbf3733c24bf2e151bca31099bfb49fef19f72bab26d21cb not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.161350 4810 scope.go:117] "RemoveContainer" containerID="7018509dad704fa1f12dfe394702b01380dd65a2edb40fddaab6e5723fd4e885" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.161906 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7018509dad704fa1f12dfe394702b01380dd65a2edb40fddaab6e5723fd4e885\": container with ID starting with 7018509dad704fa1f12dfe394702b01380dd65a2edb40fddaab6e5723fd4e885 not found: ID does not exist" containerID="7018509dad704fa1f12dfe394702b01380dd65a2edb40fddaab6e5723fd4e885" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.161924 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7018509dad704fa1f12dfe394702b01380dd65a2edb40fddaab6e5723fd4e885"} err="failed to get container status \"7018509dad704fa1f12dfe394702b01380dd65a2edb40fddaab6e5723fd4e885\": rpc error: code = NotFound desc = could not find container \"7018509dad704fa1f12dfe394702b01380dd65a2edb40fddaab6e5723fd4e885\": container with ID starting with 7018509dad704fa1f12dfe394702b01380dd65a2edb40fddaab6e5723fd4e885 not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.161939 4810 scope.go:117] "RemoveContainer" containerID="11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.162148 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029\": container with ID starting with 11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029 not found: ID does not exist" containerID="11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.162165 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029"} err="failed to get container status \"11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029\": rpc error: code = NotFound desc = could not find container \"11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029\": container with ID starting with 11a33385020787b7a6d5403d5216249c987355689266654dd5720ac611acc029 not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.162178 4810 scope.go:117] "RemoveContainer" containerID="f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.178945 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8x56x"] Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.195481 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8x56x"] Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.198234 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nzjhz"] Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.208402 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqwr"] Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.221024 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqwr"] Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.228658 4810 scope.go:117] "RemoveContainer" containerID="f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.230713 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec\": container with ID starting with f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec not found: ID does not exist" containerID="f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.230815 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec"} err="failed to get container status \"f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec\": rpc error: code = NotFound desc = could not find container \"f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec\": container with ID starting with f74d063789dd0e012adadd9de274386fda1ff14906908fdc21008af594b29eec not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.230908 4810 scope.go:117] "RemoveContainer" containerID="4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.257370 4810 scope.go:117] "RemoveContainer" containerID="5ddeecc1635379a06d57433c5a2a2770697b1e81c85508cd3e5d359778c00df0" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.275730 4810 scope.go:117] "RemoveContainer" containerID="3a66a215df1ea740c0c89f0c95bde8f7cb9d2acb79d8161a6eb47f8ffc1d2c1e" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.304907 4810 scope.go:117] "RemoveContainer" containerID="4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.305436 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750\": container with ID starting with 4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750 not found: ID does not exist" containerID="4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.305482 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750"} err="failed to get container status \"4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750\": rpc error: code = NotFound desc = could not find container \"4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750\": container with ID starting with 4bcc9dee2aca9229251eff94c8e350ccc3500f15586b6c058a14a2706996b750 not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.305514 4810 scope.go:117] "RemoveContainer" containerID="5ddeecc1635379a06d57433c5a2a2770697b1e81c85508cd3e5d359778c00df0" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.305779 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ddeecc1635379a06d57433c5a2a2770697b1e81c85508cd3e5d359778c00df0\": container with ID starting with 5ddeecc1635379a06d57433c5a2a2770697b1e81c85508cd3e5d359778c00df0 not found: ID does not exist" containerID="5ddeecc1635379a06d57433c5a2a2770697b1e81c85508cd3e5d359778c00df0" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.305801 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ddeecc1635379a06d57433c5a2a2770697b1e81c85508cd3e5d359778c00df0"} err="failed to get container status \"5ddeecc1635379a06d57433c5a2a2770697b1e81c85508cd3e5d359778c00df0\": rpc error: code = NotFound desc = could not find container \"5ddeecc1635379a06d57433c5a2a2770697b1e81c85508cd3e5d359778c00df0\": container with ID starting with 5ddeecc1635379a06d57433c5a2a2770697b1e81c85508cd3e5d359778c00df0 not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.305816 4810 scope.go:117] "RemoveContainer" containerID="3a66a215df1ea740c0c89f0c95bde8f7cb9d2acb79d8161a6eb47f8ffc1d2c1e" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.305982 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a66a215df1ea740c0c89f0c95bde8f7cb9d2acb79d8161a6eb47f8ffc1d2c1e\": container with ID starting with 3a66a215df1ea740c0c89f0c95bde8f7cb9d2acb79d8161a6eb47f8ffc1d2c1e not found: ID does not exist" containerID="3a66a215df1ea740c0c89f0c95bde8f7cb9d2acb79d8161a6eb47f8ffc1d2c1e" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.306003 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a66a215df1ea740c0c89f0c95bde8f7cb9d2acb79d8161a6eb47f8ffc1d2c1e"} err="failed to get container status \"3a66a215df1ea740c0c89f0c95bde8f7cb9d2acb79d8161a6eb47f8ffc1d2c1e\": rpc error: code = NotFound desc = could not find container \"3a66a215df1ea740c0c89f0c95bde8f7cb9d2acb79d8161a6eb47f8ffc1d2c1e\": container with ID starting with 3a66a215df1ea740c0c89f0c95bde8f7cb9d2acb79d8161a6eb47f8ffc1d2c1e not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.306018 4810 scope.go:117] "RemoveContainer" containerID="802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.328292 4810 scope.go:117] "RemoveContainer" containerID="80e68ade2286c5751793707ae43ac92711e2eb74f4626ac20edfa555d40265a7" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.365915 4810 scope.go:117] "RemoveContainer" containerID="32c465e63ce961f8d3c7edd7f41e046e00779a7ce4f911f0b001f2c539f23d33" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.382821 4810 scope.go:117] "RemoveContainer" containerID="802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.386906 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021\": container with ID starting with 802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021 not found: ID does not exist" containerID="802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.386985 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021"} err="failed to get container status \"802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021\": rpc error: code = NotFound desc = could not find container \"802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021\": container with ID starting with 802ed840cd3a515f750e169402f0072e458c50ccb4b07f1f42a636a2f2361021 not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.387049 4810 scope.go:117] "RemoveContainer" containerID="80e68ade2286c5751793707ae43ac92711e2eb74f4626ac20edfa555d40265a7" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.390006 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80e68ade2286c5751793707ae43ac92711e2eb74f4626ac20edfa555d40265a7\": container with ID starting with 80e68ade2286c5751793707ae43ac92711e2eb74f4626ac20edfa555d40265a7 not found: ID does not exist" containerID="80e68ade2286c5751793707ae43ac92711e2eb74f4626ac20edfa555d40265a7" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.390112 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80e68ade2286c5751793707ae43ac92711e2eb74f4626ac20edfa555d40265a7"} err="failed to get container status \"80e68ade2286c5751793707ae43ac92711e2eb74f4626ac20edfa555d40265a7\": rpc error: code = NotFound desc = could not find container \"80e68ade2286c5751793707ae43ac92711e2eb74f4626ac20edfa555d40265a7\": container with ID starting with 80e68ade2286c5751793707ae43ac92711e2eb74f4626ac20edfa555d40265a7 not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.390194 4810 scope.go:117] "RemoveContainer" containerID="32c465e63ce961f8d3c7edd7f41e046e00779a7ce4f911f0b001f2c539f23d33" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.392727 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32c465e63ce961f8d3c7edd7f41e046e00779a7ce4f911f0b001f2c539f23d33\": container with ID starting with 32c465e63ce961f8d3c7edd7f41e046e00779a7ce4f911f0b001f2c539f23d33 not found: ID does not exist" containerID="32c465e63ce961f8d3c7edd7f41e046e00779a7ce4f911f0b001f2c539f23d33" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.392781 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32c465e63ce961f8d3c7edd7f41e046e00779a7ce4f911f0b001f2c539f23d33"} err="failed to get container status \"32c465e63ce961f8d3c7edd7f41e046e00779a7ce4f911f0b001f2c539f23d33\": rpc error: code = NotFound desc = could not find container \"32c465e63ce961f8d3c7edd7f41e046e00779a7ce4f911f0b001f2c539f23d33\": container with ID starting with 32c465e63ce961f8d3c7edd7f41e046e00779a7ce4f911f0b001f2c539f23d33 not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.392817 4810 scope.go:117] "RemoveContainer" containerID="d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.420931 4810 scope.go:117] "RemoveContainer" containerID="55939f1fb00cd55a38e93707977d076e5eb09b6b4c8d8aa364b3532f3f893643" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.453781 4810 scope.go:117] "RemoveContainer" containerID="260043361c1ef49108c4043fd614f87158b44615189c735fe31bcf08aa862cb6" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.470537 4810 scope.go:117] "RemoveContainer" containerID="d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.471118 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101\": container with ID starting with d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101 not found: ID does not exist" containerID="d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.471148 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101"} err="failed to get container status \"d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101\": rpc error: code = NotFound desc = could not find container \"d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101\": container with ID starting with d83b9f73c9da07000e9006a377e54321a95cc43efd311a1f1981a0510da1f101 not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.471173 4810 scope.go:117] "RemoveContainer" containerID="55939f1fb00cd55a38e93707977d076e5eb09b6b4c8d8aa364b3532f3f893643" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.471499 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55939f1fb00cd55a38e93707977d076e5eb09b6b4c8d8aa364b3532f3f893643\": container with ID starting with 55939f1fb00cd55a38e93707977d076e5eb09b6b4c8d8aa364b3532f3f893643 not found: ID does not exist" containerID="55939f1fb00cd55a38e93707977d076e5eb09b6b4c8d8aa364b3532f3f893643" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.471525 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55939f1fb00cd55a38e93707977d076e5eb09b6b4c8d8aa364b3532f3f893643"} err="failed to get container status \"55939f1fb00cd55a38e93707977d076e5eb09b6b4c8d8aa364b3532f3f893643\": rpc error: code = NotFound desc = could not find container \"55939f1fb00cd55a38e93707977d076e5eb09b6b4c8d8aa364b3532f3f893643\": container with ID starting with 55939f1fb00cd55a38e93707977d076e5eb09b6b4c8d8aa364b3532f3f893643 not found: ID does not exist" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.471558 4810 scope.go:117] "RemoveContainer" containerID="260043361c1ef49108c4043fd614f87158b44615189c735fe31bcf08aa862cb6" Sep 30 08:06:48 crc kubenswrapper[4810]: E0930 08:06:48.471859 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"260043361c1ef49108c4043fd614f87158b44615189c735fe31bcf08aa862cb6\": container with ID starting with 260043361c1ef49108c4043fd614f87158b44615189c735fe31bcf08aa862cb6 not found: ID does not exist" containerID="260043361c1ef49108c4043fd614f87158b44615189c735fe31bcf08aa862cb6" Sep 30 08:06:48 crc kubenswrapper[4810]: I0930 08:06:48.471879 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"260043361c1ef49108c4043fd614f87158b44615189c735fe31bcf08aa862cb6"} err="failed to get container status \"260043361c1ef49108c4043fd614f87158b44615189c735fe31bcf08aa862cb6\": rpc error: code = NotFound desc = could not find container \"260043361c1ef49108c4043fd614f87158b44615189c735fe31bcf08aa862cb6\": container with ID starting with 260043361c1ef49108c4043fd614f87158b44615189c735fe31bcf08aa862cb6 not found: ID does not exist" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.062192 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-txqd2" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.218097 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-skf7g"] Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.219162 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015f2a85-90c3-4bd9-b77a-63209e20cc74" containerName="extract-utilities" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.219306 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="015f2a85-90c3-4bd9-b77a-63209e20cc74" containerName="extract-utilities" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.219407 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e049bda6-e38f-4357-926e-d8b767cf5ed9" containerName="marketplace-operator" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.219498 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e049bda6-e38f-4357-926e-d8b767cf5ed9" containerName="marketplace-operator" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.219605 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.219666 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.219721 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" containerName="extract-utilities" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.219773 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" containerName="extract-utilities" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.219883 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" containerName="extract-utilities" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.219946 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" containerName="extract-utilities" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.220003 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" containerName="extract-content" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.220093 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" containerName="extract-content" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.220157 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" containerName="extract-utilities" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.220210 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" containerName="extract-utilities" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.220341 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.220423 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.220497 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" containerName="extract-content" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.220556 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" containerName="extract-content" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.220614 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.220698 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.220802 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015f2a85-90c3-4bd9-b77a-63209e20cc74" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.220864 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="015f2a85-90c3-4bd9-b77a-63209e20cc74" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.220920 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" containerName="extract-content" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.220971 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" containerName="extract-content" Sep 30 08:06:49 crc kubenswrapper[4810]: E0930 08:06:49.221086 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015f2a85-90c3-4bd9-b77a-63209e20cc74" containerName="extract-content" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.221244 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="015f2a85-90c3-4bd9-b77a-63209e20cc74" containerName="extract-content" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.221456 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e049bda6-e38f-4357-926e-d8b767cf5ed9" containerName="marketplace-operator" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.221529 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.221719 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="015f2a85-90c3-4bd9-b77a-63209e20cc74" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.221903 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.221972 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" containerName="registry-server" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.222765 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.224989 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.232474 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-skf7g"] Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.322499 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="015f2a85-90c3-4bd9-b77a-63209e20cc74" path="/var/lib/kubelet/pods/015f2a85-90c3-4bd9-b77a-63209e20cc74/volumes" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.324008 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51ccfc9f-81fc-4c77-95bb-fdb5e2239829" path="/var/lib/kubelet/pods/51ccfc9f-81fc-4c77-95bb-fdb5e2239829/volumes" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.325162 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e049bda6-e38f-4357-926e-d8b767cf5ed9" path="/var/lib/kubelet/pods/e049bda6-e38f-4357-926e-d8b767cf5ed9/volumes" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.326777 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0546fe85-ddc6-4afa-afce-8ce05abba16b-utilities\") pod \"certified-operators-skf7g\" (UID: \"0546fe85-ddc6-4afa-afce-8ce05abba16b\") " pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.326899 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0546fe85-ddc6-4afa-afce-8ce05abba16b-catalog-content\") pod \"certified-operators-skf7g\" (UID: \"0546fe85-ddc6-4afa-afce-8ce05abba16b\") " pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.326930 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4mc6\" (UniqueName: \"kubernetes.io/projected/0546fe85-ddc6-4afa-afce-8ce05abba16b-kube-api-access-n4mc6\") pod \"certified-operators-skf7g\" (UID: \"0546fe85-ddc6-4afa-afce-8ce05abba16b\") " pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.327126 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb1baa89-d430-4243-9b2d-2d2932c1abe4" path="/var/lib/kubelet/pods/fb1baa89-d430-4243-9b2d-2d2932c1abe4/volumes" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.328142 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fee7bb01-669e-49de-b6fa-7cf3951fdfc2" path="/var/lib/kubelet/pods/fee7bb01-669e-49de-b6fa-7cf3951fdfc2/volumes" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.415535 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zvbp7"] Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.416649 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.422347 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.427615 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0546fe85-ddc6-4afa-afce-8ce05abba16b-utilities\") pod \"certified-operators-skf7g\" (UID: \"0546fe85-ddc6-4afa-afce-8ce05abba16b\") " pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.427832 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0546fe85-ddc6-4afa-afce-8ce05abba16b-catalog-content\") pod \"certified-operators-skf7g\" (UID: \"0546fe85-ddc6-4afa-afce-8ce05abba16b\") " pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.427906 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4mc6\" (UniqueName: \"kubernetes.io/projected/0546fe85-ddc6-4afa-afce-8ce05abba16b-kube-api-access-n4mc6\") pod \"certified-operators-skf7g\" (UID: \"0546fe85-ddc6-4afa-afce-8ce05abba16b\") " pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.429384 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0546fe85-ddc6-4afa-afce-8ce05abba16b-utilities\") pod \"certified-operators-skf7g\" (UID: \"0546fe85-ddc6-4afa-afce-8ce05abba16b\") " pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.429670 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0546fe85-ddc6-4afa-afce-8ce05abba16b-catalog-content\") pod \"certified-operators-skf7g\" (UID: \"0546fe85-ddc6-4afa-afce-8ce05abba16b\") " pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.431940 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvbp7"] Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.456548 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4mc6\" (UniqueName: \"kubernetes.io/projected/0546fe85-ddc6-4afa-afce-8ce05abba16b-kube-api-access-n4mc6\") pod \"certified-operators-skf7g\" (UID: \"0546fe85-ddc6-4afa-afce-8ce05abba16b\") " pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.529077 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-catalog-content\") pod \"redhat-marketplace-zvbp7\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.529141 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdtt2\" (UniqueName: \"kubernetes.io/projected/4ae8468f-2d42-4b13-9dbe-f8558df495cc-kube-api-access-jdtt2\") pod \"redhat-marketplace-zvbp7\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.529170 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-utilities\") pod \"redhat-marketplace-zvbp7\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.558663 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.631628 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-catalog-content\") pod \"redhat-marketplace-zvbp7\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.632185 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdtt2\" (UniqueName: \"kubernetes.io/projected/4ae8468f-2d42-4b13-9dbe-f8558df495cc-kube-api-access-jdtt2\") pod \"redhat-marketplace-zvbp7\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.632208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-catalog-content\") pod \"redhat-marketplace-zvbp7\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.632248 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-utilities\") pod \"redhat-marketplace-zvbp7\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.633078 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-utilities\") pod \"redhat-marketplace-zvbp7\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.660702 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdtt2\" (UniqueName: \"kubernetes.io/projected/4ae8468f-2d42-4b13-9dbe-f8558df495cc-kube-api-access-jdtt2\") pod \"redhat-marketplace-zvbp7\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.736086 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:49 crc kubenswrapper[4810]: I0930 08:06:49.975942 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-skf7g"] Sep 30 08:06:49 crc kubenswrapper[4810]: W0930 08:06:49.985536 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0546fe85_ddc6_4afa_afce_8ce05abba16b.slice/crio-4df58c9abbf4ee9c91caf1c5e139d022f9cba4307c5e5be84b86bde0b9373853 WatchSource:0}: Error finding container 4df58c9abbf4ee9c91caf1c5e139d022f9cba4307c5e5be84b86bde0b9373853: Status 404 returned error can't find the container with id 4df58c9abbf4ee9c91caf1c5e139d022f9cba4307c5e5be84b86bde0b9373853 Sep 30 08:06:50 crc kubenswrapper[4810]: I0930 08:06:50.068018 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skf7g" event={"ID":"0546fe85-ddc6-4afa-afce-8ce05abba16b","Type":"ContainerStarted","Data":"4df58c9abbf4ee9c91caf1c5e139d022f9cba4307c5e5be84b86bde0b9373853"} Sep 30 08:06:50 crc kubenswrapper[4810]: I0930 08:06:50.144746 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvbp7"] Sep 30 08:06:50 crc kubenswrapper[4810]: W0930 08:06:50.187231 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ae8468f_2d42_4b13_9dbe_f8558df495cc.slice/crio-f849f484610cac87045cdfbded50bf106ff96e025879b4b97ca1c8d7aa206f09 WatchSource:0}: Error finding container f849f484610cac87045cdfbded50bf106ff96e025879b4b97ca1c8d7aa206f09: Status 404 returned error can't find the container with id f849f484610cac87045cdfbded50bf106ff96e025879b4b97ca1c8d7aa206f09 Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.076060 4810 generic.go:334] "Generic (PLEG): container finished" podID="0546fe85-ddc6-4afa-afce-8ce05abba16b" containerID="2ac087f7856a7aabcbed4847a27083f8801c1075b3099443c4c438f77decdc6a" exitCode=0 Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.076165 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skf7g" event={"ID":"0546fe85-ddc6-4afa-afce-8ce05abba16b","Type":"ContainerDied","Data":"2ac087f7856a7aabcbed4847a27083f8801c1075b3099443c4c438f77decdc6a"} Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.078940 4810 generic.go:334] "Generic (PLEG): container finished" podID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" containerID="eb197027db1f95d8ede5d6c85dac7a4ac67c1014b67d6a1ca3bdc9864a3002ec" exitCode=0 Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.078997 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvbp7" event={"ID":"4ae8468f-2d42-4b13-9dbe-f8558df495cc","Type":"ContainerDied","Data":"eb197027db1f95d8ede5d6c85dac7a4ac67c1014b67d6a1ca3bdc9864a3002ec"} Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.079052 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvbp7" event={"ID":"4ae8468f-2d42-4b13-9dbe-f8558df495cc","Type":"ContainerStarted","Data":"f849f484610cac87045cdfbded50bf106ff96e025879b4b97ca1c8d7aa206f09"} Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.622692 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n6q2z"] Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.624829 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.628416 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.641696 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6q2z"] Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.674354 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-catalog-content\") pod \"redhat-operators-n6q2z\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.674437 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fknvd\" (UniqueName: \"kubernetes.io/projected/dad7c717-3a00-42d8-b71c-d2fb42083e83-kube-api-access-fknvd\") pod \"redhat-operators-n6q2z\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.674481 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-utilities\") pod \"redhat-operators-n6q2z\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.775718 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-catalog-content\") pod \"redhat-operators-n6q2z\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.775773 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fknvd\" (UniqueName: \"kubernetes.io/projected/dad7c717-3a00-42d8-b71c-d2fb42083e83-kube-api-access-fknvd\") pod \"redhat-operators-n6q2z\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.775808 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-utilities\") pod \"redhat-operators-n6q2z\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.776648 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-catalog-content\") pod \"redhat-operators-n6q2z\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.776736 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-utilities\") pod \"redhat-operators-n6q2z\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.807761 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fknvd\" (UniqueName: \"kubernetes.io/projected/dad7c717-3a00-42d8-b71c-d2fb42083e83-kube-api-access-fknvd\") pod \"redhat-operators-n6q2z\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.820665 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2lx8v"] Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.823222 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.827405 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.828603 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2lx8v"] Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.877287 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-utilities\") pod \"community-operators-2lx8v\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.877345 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-catalog-content\") pod \"community-operators-2lx8v\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.877405 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbqd8\" (UniqueName: \"kubernetes.io/projected/6596bb1a-28f7-453f-8b30-ade04d3180b0-kube-api-access-zbqd8\") pod \"community-operators-2lx8v\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.978576 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-catalog-content\") pod \"community-operators-2lx8v\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.978644 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbqd8\" (UniqueName: \"kubernetes.io/projected/6596bb1a-28f7-453f-8b30-ade04d3180b0-kube-api-access-zbqd8\") pod \"community-operators-2lx8v\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.978724 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-utilities\") pod \"community-operators-2lx8v\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.979180 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-catalog-content\") pod \"community-operators-2lx8v\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.979205 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-utilities\") pod \"community-operators-2lx8v\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:06:51 crc kubenswrapper[4810]: I0930 08:06:51.994317 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbqd8\" (UniqueName: \"kubernetes.io/projected/6596bb1a-28f7-453f-8b30-ade04d3180b0-kube-api-access-zbqd8\") pod \"community-operators-2lx8v\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:06:52 crc kubenswrapper[4810]: I0930 08:06:52.018241 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:06:52 crc kubenswrapper[4810]: I0930 08:06:52.084109 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skf7g" event={"ID":"0546fe85-ddc6-4afa-afce-8ce05abba16b","Type":"ContainerStarted","Data":"f36d28a61c6bff1bda7d15549a9f16e6ae7727c45e4efb2b0a49bed7d12878cd"} Sep 30 08:06:52 crc kubenswrapper[4810]: I0930 08:06:52.153619 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:06:52 crc kubenswrapper[4810]: I0930 08:06:52.244603 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6q2z"] Sep 30 08:06:52 crc kubenswrapper[4810]: W0930 08:06:52.253406 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddad7c717_3a00_42d8_b71c_d2fb42083e83.slice/crio-1f8bad1017a85035c6ed1696eba78a22a2f64e65cad1feecd243bd28c4ba880f WatchSource:0}: Error finding container 1f8bad1017a85035c6ed1696eba78a22a2f64e65cad1feecd243bd28c4ba880f: Status 404 returned error can't find the container with id 1f8bad1017a85035c6ed1696eba78a22a2f64e65cad1feecd243bd28c4ba880f Sep 30 08:06:52 crc kubenswrapper[4810]: I0930 08:06:52.344019 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2lx8v"] Sep 30 08:06:52 crc kubenswrapper[4810]: W0930 08:06:52.392864 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6596bb1a_28f7_453f_8b30_ade04d3180b0.slice/crio-81c4f8c2b311128c7643e911bd8f85d80cfe407b56fea94d5ffe6d305ade56d2 WatchSource:0}: Error finding container 81c4f8c2b311128c7643e911bd8f85d80cfe407b56fea94d5ffe6d305ade56d2: Status 404 returned error can't find the container with id 81c4f8c2b311128c7643e911bd8f85d80cfe407b56fea94d5ffe6d305ade56d2 Sep 30 08:06:53 crc kubenswrapper[4810]: I0930 08:06:53.091412 4810 generic.go:334] "Generic (PLEG): container finished" podID="6596bb1a-28f7-453f-8b30-ade04d3180b0" containerID="01c1c7a0b594e157177ef4fa261ee3ab0d78cf0c7f6071e18b59fa53e641000e" exitCode=0 Sep 30 08:06:53 crc kubenswrapper[4810]: I0930 08:06:53.091743 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lx8v" event={"ID":"6596bb1a-28f7-453f-8b30-ade04d3180b0","Type":"ContainerDied","Data":"01c1c7a0b594e157177ef4fa261ee3ab0d78cf0c7f6071e18b59fa53e641000e"} Sep 30 08:06:53 crc kubenswrapper[4810]: I0930 08:06:53.091775 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lx8v" event={"ID":"6596bb1a-28f7-453f-8b30-ade04d3180b0","Type":"ContainerStarted","Data":"81c4f8c2b311128c7643e911bd8f85d80cfe407b56fea94d5ffe6d305ade56d2"} Sep 30 08:06:53 crc kubenswrapper[4810]: I0930 08:06:53.096053 4810 generic.go:334] "Generic (PLEG): container finished" podID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" containerID="d8c48b26ee828189c3737a97b910929e5a31021340f6793e6a283f3518865358" exitCode=0 Sep 30 08:06:53 crc kubenswrapper[4810]: I0930 08:06:53.096140 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvbp7" event={"ID":"4ae8468f-2d42-4b13-9dbe-f8558df495cc","Type":"ContainerDied","Data":"d8c48b26ee828189c3737a97b910929e5a31021340f6793e6a283f3518865358"} Sep 30 08:06:53 crc kubenswrapper[4810]: I0930 08:06:53.099085 4810 generic.go:334] "Generic (PLEG): container finished" podID="0546fe85-ddc6-4afa-afce-8ce05abba16b" containerID="f36d28a61c6bff1bda7d15549a9f16e6ae7727c45e4efb2b0a49bed7d12878cd" exitCode=0 Sep 30 08:06:53 crc kubenswrapper[4810]: I0930 08:06:53.099215 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skf7g" event={"ID":"0546fe85-ddc6-4afa-afce-8ce05abba16b","Type":"ContainerDied","Data":"f36d28a61c6bff1bda7d15549a9f16e6ae7727c45e4efb2b0a49bed7d12878cd"} Sep 30 08:06:53 crc kubenswrapper[4810]: I0930 08:06:53.099258 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skf7g" event={"ID":"0546fe85-ddc6-4afa-afce-8ce05abba16b","Type":"ContainerStarted","Data":"ba0b9c7c9cf09e17ed65d55d498cde61fe5f4454420016a9596ff1eefce9c41d"} Sep 30 08:06:53 crc kubenswrapper[4810]: I0930 08:06:53.101237 4810 generic.go:334] "Generic (PLEG): container finished" podID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerID="2fbd121e378b38d23500b5250a8b553ae9c55b56bdae973cf1503c04e10b513d" exitCode=0 Sep 30 08:06:53 crc kubenswrapper[4810]: I0930 08:06:53.101295 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6q2z" event={"ID":"dad7c717-3a00-42d8-b71c-d2fb42083e83","Type":"ContainerDied","Data":"2fbd121e378b38d23500b5250a8b553ae9c55b56bdae973cf1503c04e10b513d"} Sep 30 08:06:53 crc kubenswrapper[4810]: I0930 08:06:53.101323 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6q2z" event={"ID":"dad7c717-3a00-42d8-b71c-d2fb42083e83","Type":"ContainerStarted","Data":"1f8bad1017a85035c6ed1696eba78a22a2f64e65cad1feecd243bd28c4ba880f"} Sep 30 08:06:54 crc kubenswrapper[4810]: I0930 08:06:54.109663 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvbp7" event={"ID":"4ae8468f-2d42-4b13-9dbe-f8558df495cc","Type":"ContainerStarted","Data":"c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6"} Sep 30 08:06:54 crc kubenswrapper[4810]: I0930 08:06:54.134392 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-skf7g" podStartSLOduration=3.6104524700000002 podStartE2EDuration="5.134374233s" podCreationTimestamp="2025-09-30 08:06:49 +0000 UTC" firstStartedPulling="2025-09-30 08:06:51.080099754 +0000 UTC m=+234.532299031" lastFinishedPulling="2025-09-30 08:06:52.604021527 +0000 UTC m=+236.056220794" observedRunningTime="2025-09-30 08:06:53.183974057 +0000 UTC m=+236.636173334" watchObservedRunningTime="2025-09-30 08:06:54.134374233 +0000 UTC m=+237.586573500" Sep 30 08:06:54 crc kubenswrapper[4810]: I0930 08:06:54.136861 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zvbp7" podStartSLOduration=2.467309403 podStartE2EDuration="5.136849855s" podCreationTimestamp="2025-09-30 08:06:49 +0000 UTC" firstStartedPulling="2025-09-30 08:06:51.080224088 +0000 UTC m=+234.532423385" lastFinishedPulling="2025-09-30 08:06:53.74976456 +0000 UTC m=+237.201963837" observedRunningTime="2025-09-30 08:06:54.132334126 +0000 UTC m=+237.584533383" watchObservedRunningTime="2025-09-30 08:06:54.136849855 +0000 UTC m=+237.589049122" Sep 30 08:06:56 crc kubenswrapper[4810]: I0930 08:06:56.123108 4810 generic.go:334] "Generic (PLEG): container finished" podID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerID="b44c192827634d889a6685c8a4a8cca5a58be2984c28befb03c4924e9f1921db" exitCode=0 Sep 30 08:06:56 crc kubenswrapper[4810]: I0930 08:06:56.123854 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6q2z" event={"ID":"dad7c717-3a00-42d8-b71c-d2fb42083e83","Type":"ContainerDied","Data":"b44c192827634d889a6685c8a4a8cca5a58be2984c28befb03c4924e9f1921db"} Sep 30 08:06:56 crc kubenswrapper[4810]: I0930 08:06:56.126375 4810 generic.go:334] "Generic (PLEG): container finished" podID="6596bb1a-28f7-453f-8b30-ade04d3180b0" containerID="4127fa66be8be0eca71c51c0a3d2849ecb0f83fa49390811d02f2d3cc19e8f1f" exitCode=0 Sep 30 08:06:56 crc kubenswrapper[4810]: I0930 08:06:56.126473 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lx8v" event={"ID":"6596bb1a-28f7-453f-8b30-ade04d3180b0","Type":"ContainerDied","Data":"4127fa66be8be0eca71c51c0a3d2849ecb0f83fa49390811d02f2d3cc19e8f1f"} Sep 30 08:06:57 crc kubenswrapper[4810]: I0930 08:06:57.137341 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6q2z" event={"ID":"dad7c717-3a00-42d8-b71c-d2fb42083e83","Type":"ContainerStarted","Data":"b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c"} Sep 30 08:06:57 crc kubenswrapper[4810]: I0930 08:06:57.140703 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lx8v" event={"ID":"6596bb1a-28f7-453f-8b30-ade04d3180b0","Type":"ContainerStarted","Data":"57d19542556ac781b42a39a2c56d445960c973887d372238c996b162c2889992"} Sep 30 08:06:57 crc kubenswrapper[4810]: I0930 08:06:57.164645 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n6q2z" podStartSLOduration=2.668240474 podStartE2EDuration="6.16460169s" podCreationTimestamp="2025-09-30 08:06:51 +0000 UTC" firstStartedPulling="2025-09-30 08:06:53.103133513 +0000 UTC m=+236.555332830" lastFinishedPulling="2025-09-30 08:06:56.599494749 +0000 UTC m=+240.051694046" observedRunningTime="2025-09-30 08:06:57.159645146 +0000 UTC m=+240.611844423" watchObservedRunningTime="2025-09-30 08:06:57.16460169 +0000 UTC m=+240.616800967" Sep 30 08:06:59 crc kubenswrapper[4810]: I0930 08:06:59.559672 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:59 crc kubenswrapper[4810]: I0930 08:06:59.560299 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:59 crc kubenswrapper[4810]: I0930 08:06:59.617211 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:06:59 crc kubenswrapper[4810]: I0930 08:06:59.649440 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2lx8v" podStartSLOduration=5.044662797 podStartE2EDuration="8.649411354s" podCreationTimestamp="2025-09-30 08:06:51 +0000 UTC" firstStartedPulling="2025-09-30 08:06:53.093472505 +0000 UTC m=+236.545671782" lastFinishedPulling="2025-09-30 08:06:56.698221072 +0000 UTC m=+240.150420339" observedRunningTime="2025-09-30 08:06:57.18586584 +0000 UTC m=+240.638065117" watchObservedRunningTime="2025-09-30 08:06:59.649411354 +0000 UTC m=+243.101610661" Sep 30 08:06:59 crc kubenswrapper[4810]: I0930 08:06:59.736628 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:59 crc kubenswrapper[4810]: I0930 08:06:59.736724 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:06:59 crc kubenswrapper[4810]: I0930 08:06:59.789800 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:07:00 crc kubenswrapper[4810]: I0930 08:07:00.220041 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-skf7g" Sep 30 08:07:00 crc kubenswrapper[4810]: I0930 08:07:00.222716 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 08:07:02 crc kubenswrapper[4810]: I0930 08:07:02.018704 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:07:02 crc kubenswrapper[4810]: I0930 08:07:02.018765 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:07:02 crc kubenswrapper[4810]: I0930 08:07:02.071305 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:07:02 crc kubenswrapper[4810]: I0930 08:07:02.154650 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:07:02 crc kubenswrapper[4810]: I0930 08:07:02.154891 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:07:02 crc kubenswrapper[4810]: I0930 08:07:02.214043 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:07:02 crc kubenswrapper[4810]: I0930 08:07:02.218452 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:07:03 crc kubenswrapper[4810]: I0930 08:07:03.243072 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.235144 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" podUID="95f2618f-a0a5-43a2-b639-25944e913395" containerName="oauth-openshift" containerID="cri-o://aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa" gracePeriod=15 Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.685151 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.738526 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2"] Sep 30 08:07:13 crc kubenswrapper[4810]: E0930 08:07:13.739379 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f2618f-a0a5-43a2-b639-25944e913395" containerName="oauth-openshift" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.739420 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f2618f-a0a5-43a2-b639-25944e913395" containerName="oauth-openshift" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.739693 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f2618f-a0a5-43a2-b639-25944e913395" containerName="oauth-openshift" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.741514 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.749477 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2"] Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.866452 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/95f2618f-a0a5-43a2-b639-25944e913395-audit-dir\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.866607 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-idp-0-file-data\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.866663 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-router-certs\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.866699 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-login\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.866771 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-provider-selection\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.866821 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-cliconfig\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.866883 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-error\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.866922 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-ocp-branding-template\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.866998 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-session\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867056 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-serving-cert\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867154 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-service-ca\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867191 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-audit-policies\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867228 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc4cz\" (UniqueName: \"kubernetes.io/projected/95f2618f-a0a5-43a2-b639-25944e913395-kube-api-access-zc4cz\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867309 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-trusted-ca-bundle\") pod \"95f2618f-a0a5-43a2-b639-25944e913395\" (UID: \"95f2618f-a0a5-43a2-b639-25944e913395\") " Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867597 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867681 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-template-login\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867733 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867785 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-audit-policies\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867865 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867899 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-session\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.867941 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.868030 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.868079 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-template-error\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.868147 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.868178 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.868197 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.868347 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.868399 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxb4c\" (UniqueName: \"kubernetes.io/projected/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-kube-api-access-hxb4c\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.868512 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-audit-dir\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.868516 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95f2618f-a0a5-43a2-b639-25944e913395-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.868535 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.868779 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.869135 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.869379 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.887656 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f2618f-a0a5-43a2-b639-25944e913395-kube-api-access-zc4cz" (OuterVolumeSpecName: "kube-api-access-zc4cz") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "kube-api-access-zc4cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.887877 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.887953 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.888482 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.889056 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.889670 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.889864 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.890352 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.890445 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "95f2618f-a0a5-43a2-b639-25944e913395" (UID: "95f2618f-a0a5-43a2-b639-25944e913395"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.969694 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.969774 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-template-login\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.969819 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.969871 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-audit-policies\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.970102 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-session\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.971321 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-audit-policies\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.970138 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.971616 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.972540 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.973131 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.973346 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.973387 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-template-error\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.973472 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.973659 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.973720 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974231 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxb4c\" (UniqueName: \"kubernetes.io/projected/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-kube-api-access-hxb4c\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974381 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-audit-dir\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974478 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-audit-dir\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974652 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974677 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974698 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974718 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974737 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974755 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974776 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974794 4810 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974811 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc4cz\" (UniqueName: \"kubernetes.io/projected/95f2618f-a0a5-43a2-b639-25944e913395-kube-api-access-zc4cz\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974829 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974849 4810 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/95f2618f-a0a5-43a2-b639-25944e913395-audit-dir\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974867 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.974938 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/95f2618f-a0a5-43a2-b639-25944e913395-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.975158 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.975486 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-session\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.977074 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.977605 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.977840 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-template-login\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.979899 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-user-template-error\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.980131 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.980542 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:13 crc kubenswrapper[4810]: I0930 08:07:13.981076 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.003115 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxb4c\" (UniqueName: \"kubernetes.io/projected/5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1-kube-api-access-hxb4c\") pod \"oauth-openshift-6b8bc975dc-gxwr2\" (UID: \"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1\") " pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.063461 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.254629 4810 generic.go:334] "Generic (PLEG): container finished" podID="95f2618f-a0a5-43a2-b639-25944e913395" containerID="aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa" exitCode=0 Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.254689 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" event={"ID":"95f2618f-a0a5-43a2-b639-25944e913395","Type":"ContainerDied","Data":"aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa"} Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.254727 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.254770 4810 scope.go:117] "RemoveContainer" containerID="aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa" Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.254734 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nzjhz" event={"ID":"95f2618f-a0a5-43a2-b639-25944e913395","Type":"ContainerDied","Data":"996c3b09d20d02362147173fff81f2b0299802853d5835a16b719dcf96889e34"} Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.289763 4810 scope.go:117] "RemoveContainer" containerID="aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa" Sep 30 08:07:14 crc kubenswrapper[4810]: E0930 08:07:14.290624 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa\": container with ID starting with aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa not found: ID does not exist" containerID="aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa" Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.290725 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa"} err="failed to get container status \"aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa\": rpc error: code = NotFound desc = could not find container \"aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa\": container with ID starting with aab32e9b29a9ff01b3a5896917067ac4883fd14761a37ca0031f6914a92848fa not found: ID does not exist" Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.309322 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nzjhz"] Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.315923 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nzjhz"] Sep 30 08:07:14 crc kubenswrapper[4810]: I0930 08:07:14.390559 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2"] Sep 30 08:07:14 crc kubenswrapper[4810]: W0930 08:07:14.399967 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5435fd16_ff1e_4a66_9e9a_f2e1d27a8cd1.slice/crio-a37d191da7f9a13ce392a81edf5391019149503c64683921211b9c2627b1a32b WatchSource:0}: Error finding container a37d191da7f9a13ce392a81edf5391019149503c64683921211b9c2627b1a32b: Status 404 returned error can't find the container with id a37d191da7f9a13ce392a81edf5391019149503c64683921211b9c2627b1a32b Sep 30 08:07:15 crc kubenswrapper[4810]: I0930 08:07:15.264588 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" event={"ID":"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1","Type":"ContainerStarted","Data":"8ef432da0ce7ad06e4947c00e58668167278be5e846026f2a875ed155e428837"} Sep 30 08:07:15 crc kubenswrapper[4810]: I0930 08:07:15.265018 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:15 crc kubenswrapper[4810]: I0930 08:07:15.265051 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" event={"ID":"5435fd16-ff1e-4a66-9e9a-f2e1d27a8cd1","Type":"ContainerStarted","Data":"a37d191da7f9a13ce392a81edf5391019149503c64683921211b9c2627b1a32b"} Sep 30 08:07:15 crc kubenswrapper[4810]: I0930 08:07:15.274155 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" Sep 30 08:07:15 crc kubenswrapper[4810]: I0930 08:07:15.321204 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f2618f-a0a5-43a2-b639-25944e913395" path="/var/lib/kubelet/pods/95f2618f-a0a5-43a2-b639-25944e913395/volumes" Sep 30 08:07:15 crc kubenswrapper[4810]: I0930 08:07:15.350763 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6b8bc975dc-gxwr2" podStartSLOduration=27.350735079 podStartE2EDuration="27.350735079s" podCreationTimestamp="2025-09-30 08:06:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:07:15.30678859 +0000 UTC m=+258.758987977" watchObservedRunningTime="2025-09-30 08:07:15.350735079 +0000 UTC m=+258.802934376" Sep 30 08:09:15 crc kubenswrapper[4810]: I0930 08:09:15.912195 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:09:15 crc kubenswrapper[4810]: I0930 08:09:15.912827 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:09:45 crc kubenswrapper[4810]: I0930 08:09:45.912117 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:09:45 crc kubenswrapper[4810]: I0930 08:09:45.912834 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:10:15 crc kubenswrapper[4810]: I0930 08:10:15.911508 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:10:15 crc kubenswrapper[4810]: I0930 08:10:15.912171 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:10:15 crc kubenswrapper[4810]: I0930 08:10:15.912246 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:10:15 crc kubenswrapper[4810]: I0930 08:10:15.913176 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33485b79f13ae3509d05af432da899150f9b5186a42141758f7a50e6033f5705"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:10:15 crc kubenswrapper[4810]: I0930 08:10:15.913347 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://33485b79f13ae3509d05af432da899150f9b5186a42141758f7a50e6033f5705" gracePeriod=600 Sep 30 08:10:16 crc kubenswrapper[4810]: I0930 08:10:16.493182 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="33485b79f13ae3509d05af432da899150f9b5186a42141758f7a50e6033f5705" exitCode=0 Sep 30 08:10:16 crc kubenswrapper[4810]: I0930 08:10:16.493283 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"33485b79f13ae3509d05af432da899150f9b5186a42141758f7a50e6033f5705"} Sep 30 08:10:16 crc kubenswrapper[4810]: I0930 08:10:16.493382 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"3e8713aca2fba3818ce4206358c20335307a8230c5f24759e0b223332c10c94d"} Sep 30 08:10:16 crc kubenswrapper[4810]: I0930 08:10:16.493408 4810 scope.go:117] "RemoveContainer" containerID="97e5b7218c44ebaef9d110939e1a69d5034ee2c2f2c88ba94ec5480ba9669ab4" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.697509 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kjgsl"] Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.699961 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.709955 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kjgsl"] Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.848638 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4e699307-c811-431b-926c-4c6861eb2a67-bound-sa-token\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.848704 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.848731 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4e699307-c811-431b-926c-4c6861eb2a67-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.848746 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4e699307-c811-431b-926c-4c6861eb2a67-registry-certificates\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.848763 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4e699307-c811-431b-926c-4c6861eb2a67-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.848799 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e699307-c811-431b-926c-4c6861eb2a67-trusted-ca\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.848854 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4e699307-c811-431b-926c-4c6861eb2a67-registry-tls\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.848887 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5zc9\" (UniqueName: \"kubernetes.io/projected/4e699307-c811-431b-926c-4c6861eb2a67-kube-api-access-s5zc9\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.867431 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.949638 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4e699307-c811-431b-926c-4c6861eb2a67-registry-tls\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.950172 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5zc9\" (UniqueName: \"kubernetes.io/projected/4e699307-c811-431b-926c-4c6861eb2a67-kube-api-access-s5zc9\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.950325 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4e699307-c811-431b-926c-4c6861eb2a67-bound-sa-token\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.950513 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4e699307-c811-431b-926c-4c6861eb2a67-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.950654 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4e699307-c811-431b-926c-4c6861eb2a67-registry-certificates\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.950838 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4e699307-c811-431b-926c-4c6861eb2a67-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.951021 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e699307-c811-431b-926c-4c6861eb2a67-trusted-ca\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.950928 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4e699307-c811-431b-926c-4c6861eb2a67-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.952502 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e699307-c811-431b-926c-4c6861eb2a67-trusted-ca\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.952555 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4e699307-c811-431b-926c-4c6861eb2a67-registry-certificates\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.957004 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4e699307-c811-431b-926c-4c6861eb2a67-registry-tls\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.960918 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4e699307-c811-431b-926c-4c6861eb2a67-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.966808 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5zc9\" (UniqueName: \"kubernetes.io/projected/4e699307-c811-431b-926c-4c6861eb2a67-kube-api-access-s5zc9\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:42 crc kubenswrapper[4810]: I0930 08:10:42.980721 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4e699307-c811-431b-926c-4c6861eb2a67-bound-sa-token\") pod \"image-registry-66df7c8f76-kjgsl\" (UID: \"4e699307-c811-431b-926c-4c6861eb2a67\") " pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:43 crc kubenswrapper[4810]: I0930 08:10:43.043938 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:43 crc kubenswrapper[4810]: I0930 08:10:43.231651 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kjgsl"] Sep 30 08:10:43 crc kubenswrapper[4810]: I0930 08:10:43.691524 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" event={"ID":"4e699307-c811-431b-926c-4c6861eb2a67","Type":"ContainerStarted","Data":"cc3f9bda385b5caec8cd062db904928215097c273c3e5739d380edc8c41ff713"} Sep 30 08:10:43 crc kubenswrapper[4810]: I0930 08:10:43.691611 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" event={"ID":"4e699307-c811-431b-926c-4c6861eb2a67","Type":"ContainerStarted","Data":"1a8206b3c5ba069fbe9b312ba31adafcf3dc33377f92ee034b612f8288d2fde7"} Sep 30 08:10:43 crc kubenswrapper[4810]: I0930 08:10:43.691752 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:10:43 crc kubenswrapper[4810]: I0930 08:10:43.728690 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" podStartSLOduration=1.728650262 podStartE2EDuration="1.728650262s" podCreationTimestamp="2025-09-30 08:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:10:43.722876755 +0000 UTC m=+467.175076072" watchObservedRunningTime="2025-09-30 08:10:43.728650262 +0000 UTC m=+467.180849579" Sep 30 08:11:03 crc kubenswrapper[4810]: I0930 08:11:03.053596 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-kjgsl" Sep 30 08:11:03 crc kubenswrapper[4810]: I0930 08:11:03.131168 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4bvx"] Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.185657 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" podUID="4c2e4c71-7e20-4d81-8531-f673f4c9341f" containerName="registry" containerID="cri-o://6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6" gracePeriod=30 Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.591421 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.714305 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-tls\") pod \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.714370 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-certificates\") pod \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.714412 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-bound-sa-token\") pod \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.714461 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-trusted-ca\") pod \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.714524 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8sgm\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-kube-api-access-n8sgm\") pod \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.714554 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4c2e4c71-7e20-4d81-8531-f673f4c9341f-ca-trust-extracted\") pod \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.714699 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.714728 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4c2e4c71-7e20-4d81-8531-f673f4c9341f-installation-pull-secrets\") pod \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\" (UID: \"4c2e4c71-7e20-4d81-8531-f673f4c9341f\") " Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.715833 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4c2e4c71-7e20-4d81-8531-f673f4c9341f" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.717093 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4c2e4c71-7e20-4d81-8531-f673f4c9341f" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.720527 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c2e4c71-7e20-4d81-8531-f673f4c9341f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4c2e4c71-7e20-4d81-8531-f673f4c9341f" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.722907 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-kube-api-access-n8sgm" (OuterVolumeSpecName: "kube-api-access-n8sgm") pod "4c2e4c71-7e20-4d81-8531-f673f4c9341f" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f"). InnerVolumeSpecName "kube-api-access-n8sgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.723076 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4c2e4c71-7e20-4d81-8531-f673f4c9341f" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.723394 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4c2e4c71-7e20-4d81-8531-f673f4c9341f" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.744615 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c2e4c71-7e20-4d81-8531-f673f4c9341f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4c2e4c71-7e20-4d81-8531-f673f4c9341f" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.752571 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "4c2e4c71-7e20-4d81-8531-f673f4c9341f" (UID: "4c2e4c71-7e20-4d81-8531-f673f4c9341f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.816747 4810 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.816808 4810 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.816833 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.816852 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4c2e4c71-7e20-4d81-8531-f673f4c9341f-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.816871 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8sgm\" (UniqueName: \"kubernetes.io/projected/4c2e4c71-7e20-4d81-8531-f673f4c9341f-kube-api-access-n8sgm\") on node \"crc\" DevicePath \"\"" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.816888 4810 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4c2e4c71-7e20-4d81-8531-f673f4c9341f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 30 08:11:28 crc kubenswrapper[4810]: I0930 08:11:28.816907 4810 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4c2e4c71-7e20-4d81-8531-f673f4c9341f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 30 08:11:29 crc kubenswrapper[4810]: I0930 08:11:29.018245 4810 generic.go:334] "Generic (PLEG): container finished" podID="4c2e4c71-7e20-4d81-8531-f673f4c9341f" containerID="6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6" exitCode=0 Sep 30 08:11:29 crc kubenswrapper[4810]: I0930 08:11:29.018645 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" event={"ID":"4c2e4c71-7e20-4d81-8531-f673f4c9341f","Type":"ContainerDied","Data":"6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6"} Sep 30 08:11:29 crc kubenswrapper[4810]: I0930 08:11:29.018690 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" event={"ID":"4c2e4c71-7e20-4d81-8531-f673f4c9341f","Type":"ContainerDied","Data":"3ec7bd439fc03cf75c8da3e4b679811544dc48a87d854c93c7348939eaac5acd"} Sep 30 08:11:29 crc kubenswrapper[4810]: I0930 08:11:29.018717 4810 scope.go:117] "RemoveContainer" containerID="6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6" Sep 30 08:11:29 crc kubenswrapper[4810]: I0930 08:11:29.018877 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k4bvx" Sep 30 08:11:29 crc kubenswrapper[4810]: I0930 08:11:29.040576 4810 scope.go:117] "RemoveContainer" containerID="6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6" Sep 30 08:11:29 crc kubenswrapper[4810]: E0930 08:11:29.040963 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6\": container with ID starting with 6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6 not found: ID does not exist" containerID="6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6" Sep 30 08:11:29 crc kubenswrapper[4810]: I0930 08:11:29.041000 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6"} err="failed to get container status \"6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6\": rpc error: code = NotFound desc = could not find container \"6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6\": container with ID starting with 6192b4f2d4f2647655e8b19f510c1c65bc12b896f1f599a1255e3ae58c86fab6 not found: ID does not exist" Sep 30 08:11:29 crc kubenswrapper[4810]: I0930 08:11:29.057557 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4bvx"] Sep 30 08:11:29 crc kubenswrapper[4810]: I0930 08:11:29.063105 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4bvx"] Sep 30 08:11:29 crc kubenswrapper[4810]: I0930 08:11:29.319685 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c2e4c71-7e20-4d81-8531-f673f4c9341f" path="/var/lib/kubelet/pods/4c2e4c71-7e20-4d81-8531-f673f4c9341f/volumes" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.167045 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vbt4m"] Sep 30 08:12:01 crc kubenswrapper[4810]: E0930 08:12:01.168425 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2e4c71-7e20-4d81-8531-f673f4c9341f" containerName="registry" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.168516 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2e4c71-7e20-4d81-8531-f673f4c9341f" containerName="registry" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.169076 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c2e4c71-7e20-4d81-8531-f673f4c9341f" containerName="registry" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.170397 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-vbt4m" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.176798 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bdbrw"] Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.178266 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-bdbrw" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.180757 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.181804 4810 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-v769x" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.182565 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.182763 4810 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-cjjhm" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.185110 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vbt4m"] Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.188323 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bdbrw"] Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.210896 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-wdstp"] Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.212119 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-wdstp" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.214530 4810 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-9n24v" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.231414 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-wdstp"] Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.315002 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7wjv\" (UniqueName: \"kubernetes.io/projected/28b9d5cb-a422-4629-ba55-8bcd92f69971-kube-api-access-p7wjv\") pod \"cert-manager-cainjector-7f985d654d-vbt4m\" (UID: \"28b9d5cb-a422-4629-ba55-8bcd92f69971\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vbt4m" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.315044 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dgm9\" (UniqueName: \"kubernetes.io/projected/8090d644-f567-428e-a3b4-0d529d9bccb9-kube-api-access-9dgm9\") pod \"cert-manager-webhook-5655c58dd6-wdstp\" (UID: \"8090d644-f567-428e-a3b4-0d529d9bccb9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-wdstp" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.315075 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-547bp\" (UniqueName: \"kubernetes.io/projected/14ff18f7-0dac-4905-a94e-5a687ad05c8a-kube-api-access-547bp\") pod \"cert-manager-5b446d88c5-bdbrw\" (UID: \"14ff18f7-0dac-4905-a94e-5a687ad05c8a\") " pod="cert-manager/cert-manager-5b446d88c5-bdbrw" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.416709 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7wjv\" (UniqueName: \"kubernetes.io/projected/28b9d5cb-a422-4629-ba55-8bcd92f69971-kube-api-access-p7wjv\") pod \"cert-manager-cainjector-7f985d654d-vbt4m\" (UID: \"28b9d5cb-a422-4629-ba55-8bcd92f69971\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vbt4m" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.417213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dgm9\" (UniqueName: \"kubernetes.io/projected/8090d644-f567-428e-a3b4-0d529d9bccb9-kube-api-access-9dgm9\") pod \"cert-manager-webhook-5655c58dd6-wdstp\" (UID: \"8090d644-f567-428e-a3b4-0d529d9bccb9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-wdstp" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.417465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-547bp\" (UniqueName: \"kubernetes.io/projected/14ff18f7-0dac-4905-a94e-5a687ad05c8a-kube-api-access-547bp\") pod \"cert-manager-5b446d88c5-bdbrw\" (UID: \"14ff18f7-0dac-4905-a94e-5a687ad05c8a\") " pod="cert-manager/cert-manager-5b446d88c5-bdbrw" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.437120 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dgm9\" (UniqueName: \"kubernetes.io/projected/8090d644-f567-428e-a3b4-0d529d9bccb9-kube-api-access-9dgm9\") pod \"cert-manager-webhook-5655c58dd6-wdstp\" (UID: \"8090d644-f567-428e-a3b4-0d529d9bccb9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-wdstp" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.438060 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-547bp\" (UniqueName: \"kubernetes.io/projected/14ff18f7-0dac-4905-a94e-5a687ad05c8a-kube-api-access-547bp\") pod \"cert-manager-5b446d88c5-bdbrw\" (UID: \"14ff18f7-0dac-4905-a94e-5a687ad05c8a\") " pod="cert-manager/cert-manager-5b446d88c5-bdbrw" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.439577 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7wjv\" (UniqueName: \"kubernetes.io/projected/28b9d5cb-a422-4629-ba55-8bcd92f69971-kube-api-access-p7wjv\") pod \"cert-manager-cainjector-7f985d654d-vbt4m\" (UID: \"28b9d5cb-a422-4629-ba55-8bcd92f69971\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vbt4m" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.509665 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-vbt4m" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.518879 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-bdbrw" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.530481 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-wdstp" Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.800891 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bdbrw"] Sep 30 08:12:01 crc kubenswrapper[4810]: I0930 08:12:01.811240 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 08:12:02 crc kubenswrapper[4810]: I0930 08:12:02.085088 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vbt4m"] Sep 30 08:12:02 crc kubenswrapper[4810]: I0930 08:12:02.093854 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-wdstp"] Sep 30 08:12:02 crc kubenswrapper[4810]: W0930 08:12:02.094087 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28b9d5cb_a422_4629_ba55_8bcd92f69971.slice/crio-0a9b73dd11e8ed1cfe361a333878733ffffc30108f7035a638e069124e0dd09c WatchSource:0}: Error finding container 0a9b73dd11e8ed1cfe361a333878733ffffc30108f7035a638e069124e0dd09c: Status 404 returned error can't find the container with id 0a9b73dd11e8ed1cfe361a333878733ffffc30108f7035a638e069124e0dd09c Sep 30 08:12:02 crc kubenswrapper[4810]: W0930 08:12:02.106690 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8090d644_f567_428e_a3b4_0d529d9bccb9.slice/crio-21d3234c1d7c5daa9e3daa123486dc0a2bdb395c8b81994ab7fa2847b29d4afe WatchSource:0}: Error finding container 21d3234c1d7c5daa9e3daa123486dc0a2bdb395c8b81994ab7fa2847b29d4afe: Status 404 returned error can't find the container with id 21d3234c1d7c5daa9e3daa123486dc0a2bdb395c8b81994ab7fa2847b29d4afe Sep 30 08:12:02 crc kubenswrapper[4810]: I0930 08:12:02.253257 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-wdstp" event={"ID":"8090d644-f567-428e-a3b4-0d529d9bccb9","Type":"ContainerStarted","Data":"21d3234c1d7c5daa9e3daa123486dc0a2bdb395c8b81994ab7fa2847b29d4afe"} Sep 30 08:12:02 crc kubenswrapper[4810]: I0930 08:12:02.255153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-vbt4m" event={"ID":"28b9d5cb-a422-4629-ba55-8bcd92f69971","Type":"ContainerStarted","Data":"0a9b73dd11e8ed1cfe361a333878733ffffc30108f7035a638e069124e0dd09c"} Sep 30 08:12:02 crc kubenswrapper[4810]: I0930 08:12:02.256172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-bdbrw" event={"ID":"14ff18f7-0dac-4905-a94e-5a687ad05c8a","Type":"ContainerStarted","Data":"496f4be79848c2e0f3d1d8ec1cd909f755e1373f2c6311ece0a9f7d17b041cef"} Sep 30 08:12:07 crc kubenswrapper[4810]: I0930 08:12:07.293255 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-wdstp" event={"ID":"8090d644-f567-428e-a3b4-0d529d9bccb9","Type":"ContainerStarted","Data":"c51f589c64c8b4ce22e099d04597f0e18fb77cabc18c38492caae064f160c876"} Sep 30 08:12:07 crc kubenswrapper[4810]: I0930 08:12:07.293947 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-wdstp" Sep 30 08:12:07 crc kubenswrapper[4810]: I0930 08:12:07.295837 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-vbt4m" event={"ID":"28b9d5cb-a422-4629-ba55-8bcd92f69971","Type":"ContainerStarted","Data":"4d8a62f1ee06e1533cae11637e617ebe03c0500ee5a4df118956110368fa3a4b"} Sep 30 08:12:07 crc kubenswrapper[4810]: I0930 08:12:07.298632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-bdbrw" event={"ID":"14ff18f7-0dac-4905-a94e-5a687ad05c8a","Type":"ContainerStarted","Data":"fd2113dc6d0843ecacaf8627c8b982c2b4db8fb73b6f412d3ce5d976b85a9d4c"} Sep 30 08:12:07 crc kubenswrapper[4810]: I0930 08:12:07.316726 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-wdstp" podStartSLOduration=2.065123015 podStartE2EDuration="6.316707596s" podCreationTimestamp="2025-09-30 08:12:01 +0000 UTC" firstStartedPulling="2025-09-30 08:12:02.110105287 +0000 UTC m=+545.562304554" lastFinishedPulling="2025-09-30 08:12:06.361689848 +0000 UTC m=+549.813889135" observedRunningTime="2025-09-30 08:12:07.31418441 +0000 UTC m=+550.766383707" watchObservedRunningTime="2025-09-30 08:12:07.316707596 +0000 UTC m=+550.768906883" Sep 30 08:12:07 crc kubenswrapper[4810]: I0930 08:12:07.356042 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-bdbrw" podStartSLOduration=1.814221772 podStartE2EDuration="6.356015163s" podCreationTimestamp="2025-09-30 08:12:01 +0000 UTC" firstStartedPulling="2025-09-30 08:12:01.810176626 +0000 UTC m=+545.262375893" lastFinishedPulling="2025-09-30 08:12:06.351970007 +0000 UTC m=+549.804169284" observedRunningTime="2025-09-30 08:12:07.351754725 +0000 UTC m=+550.803954002" watchObservedRunningTime="2025-09-30 08:12:07.356015163 +0000 UTC m=+550.808214450" Sep 30 08:12:07 crc kubenswrapper[4810]: I0930 08:12:07.412330 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-vbt4m" podStartSLOduration=2.154698367 podStartE2EDuration="6.412302238s" podCreationTimestamp="2025-09-30 08:12:01 +0000 UTC" firstStartedPulling="2025-09-30 08:12:02.09818906 +0000 UTC m=+545.550388327" lastFinishedPulling="2025-09-30 08:12:06.355792921 +0000 UTC m=+549.807992198" observedRunningTime="2025-09-30 08:12:07.407847765 +0000 UTC m=+550.860047042" watchObservedRunningTime="2025-09-30 08:12:07.412302238 +0000 UTC m=+550.864501525" Sep 30 08:12:11 crc kubenswrapper[4810]: I0930 08:12:11.536101 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-wdstp" Sep 30 08:12:11 crc kubenswrapper[4810]: I0930 08:12:11.996621 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9bkcj"] Sep 30 08:12:11 crc kubenswrapper[4810]: I0930 08:12:11.997559 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovn-controller" containerID="cri-o://9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644" gracePeriod=30 Sep 30 08:12:11 crc kubenswrapper[4810]: I0930 08:12:11.997645 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="nbdb" containerID="cri-o://03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89" gracePeriod=30 Sep 30 08:12:11 crc kubenswrapper[4810]: I0930 08:12:11.997709 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38" gracePeriod=30 Sep 30 08:12:11 crc kubenswrapper[4810]: I0930 08:12:11.997791 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="northd" containerID="cri-o://114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f" gracePeriod=30 Sep 30 08:12:11 crc kubenswrapper[4810]: I0930 08:12:11.997831 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="kube-rbac-proxy-node" containerID="cri-o://cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5" gracePeriod=30 Sep 30 08:12:11 crc kubenswrapper[4810]: I0930 08:12:11.997871 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovn-acl-logging" containerID="cri-o://5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5" gracePeriod=30 Sep 30 08:12:11 crc kubenswrapper[4810]: I0930 08:12:11.997733 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="sbdb" containerID="cri-o://62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6" gracePeriod=30 Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.058945 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" containerID="cri-o://933fa6a06cdfcff44c3ae380d64b817e01623656491d875120a01cdfbdbdff55" gracePeriod=30 Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.338874 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovnkube-controller/3.log" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.341636 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovn-acl-logging/0.log" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342097 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovn-controller/0.log" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342481 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="933fa6a06cdfcff44c3ae380d64b817e01623656491d875120a01cdfbdbdff55" exitCode=0 Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342513 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6" exitCode=0 Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342524 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89" exitCode=0 Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342534 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f" exitCode=0 Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342544 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38" exitCode=0 Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342553 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5" exitCode=0 Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342563 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5" exitCode=143 Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342571 4810 generic.go:334] "Generic (PLEG): container finished" podID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerID="9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644" exitCode=143 Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342528 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"933fa6a06cdfcff44c3ae380d64b817e01623656491d875120a01cdfbdbdff55"} Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342644 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6"} Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342664 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89"} Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342678 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f"} Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342690 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38"} Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342700 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5"} Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342711 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5"} Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342721 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644"} Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.342742 4810 scope.go:117] "RemoveContainer" containerID="08e06413e719c5c40aae1d6bcacfd2a48218ee8ddf7a9eb079c55eba4f93879b" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.346245 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n85f4_f5367714-1633-4695-af4e-4eb6419daa96/kube-multus/2.log" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.346974 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n85f4_f5367714-1633-4695-af4e-4eb6419daa96/kube-multus/1.log" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.347071 4810 generic.go:334] "Generic (PLEG): container finished" podID="f5367714-1633-4695-af4e-4eb6419daa96" containerID="8a0c05c5c70b22d510babd5c578400d035ba077b50083c564906b0e8dc68be45" exitCode=2 Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.347120 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n85f4" event={"ID":"f5367714-1633-4695-af4e-4eb6419daa96","Type":"ContainerDied","Data":"8a0c05c5c70b22d510babd5c578400d035ba077b50083c564906b0e8dc68be45"} Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.348923 4810 scope.go:117] "RemoveContainer" containerID="8a0c05c5c70b22d510babd5c578400d035ba077b50083c564906b0e8dc68be45" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.349302 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-n85f4_openshift-multus(f5367714-1633-4695-af4e-4eb6419daa96)\"" pod="openshift-multus/multus-n85f4" podUID="f5367714-1633-4695-af4e-4eb6419daa96" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.357109 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovn-acl-logging/0.log" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.357958 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovn-controller/0.log" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.358563 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.371864 4810 scope.go:117] "RemoveContainer" containerID="535e0d3c6cc6a5a6213a84bea8dc435af8af89a7ce2db5bc1b0bf9e499e08857" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440583 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5g7w2"] Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.440828 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440844 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.440853 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440859 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.440867 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440873 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.440880 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="sbdb" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440886 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="sbdb" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.440895 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440901 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.440911 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="kubecfg-setup" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440917 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="kubecfg-setup" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.440926 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovn-acl-logging" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440932 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovn-acl-logging" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.440942 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440948 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.440958 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="kube-rbac-proxy-node" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440964 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="kube-rbac-proxy-node" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.440975 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="northd" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440981 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="northd" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.440991 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovn-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.440997 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovn-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.441006 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="nbdb" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441012 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="nbdb" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441119 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="northd" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441132 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="sbdb" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441140 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="nbdb" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441150 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="kube-rbac-proxy-node" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441157 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441165 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovn-acl-logging" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441174 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441181 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441187 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441194 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovn-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: E0930 08:12:12.441294 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441301 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441387 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.441398 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" containerName="ovnkube-controller" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.443138 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.510838 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-systemd\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.510907 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.510928 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-systemd-units\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.510949 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-node-log\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.510980 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx6mf\" (UniqueName: \"kubernetes.io/projected/aacbcdc8-7ef5-473b-8055-145f40040d5d-kube-api-access-kx6mf\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511007 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-ovn\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511021 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-log-socket\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511045 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-bin\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511060 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-config\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511073 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-netns\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511089 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-ovn-kubernetes\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511103 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-netd\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511126 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-script-lib\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511143 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-env-overrides\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511175 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-var-lib-openvswitch\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511197 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-kubelet\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511219 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovn-node-metrics-cert\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511245 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-slash\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511259 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-openvswitch\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511298 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-etc-openvswitch\") pod \"aacbcdc8-7ef5-473b-8055-145f40040d5d\" (UID: \"aacbcdc8-7ef5-473b-8055-145f40040d5d\") " Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511425 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-var-lib-openvswitch\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511451 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-run-netns\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511499 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-etc-openvswitch\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511517 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-run-openvswitch\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511533 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/037365fc-5b0d-4e83-a6e8-eee1e4151047-env-overrides\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511550 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-run-ovn-kubernetes\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511551 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511645 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511580 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/037365fc-5b0d-4e83-a6e8-eee1e4151047-ovnkube-script-lib\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511768 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-cni-bin\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511844 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/037365fc-5b0d-4e83-a6e8-eee1e4151047-ovnkube-config\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511908 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-log-socket\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512075 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-node-log\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512142 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-cni-netd\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512210 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/037365fc-5b0d-4e83-a6e8-eee1e4151047-ovn-node-metrics-cert\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512355 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-run-systemd\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511675 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.511989 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512245 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512288 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512332 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512407 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512348 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-log-socket" (OuterVolumeSpecName: "log-socket") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512359 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-node-log" (OuterVolumeSpecName: "node-log") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512406 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-run-ovn\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512473 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x8cp\" (UniqueName: \"kubernetes.io/projected/037365fc-5b0d-4e83-a6e8-eee1e4151047-kube-api-access-9x8cp\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512508 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-systemd-units\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512530 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-kubelet\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512551 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-slash\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512573 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512622 4810 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512638 4810 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-systemd-units\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512649 4810 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-node-log\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512660 4810 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512670 4810 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-log-socket\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512681 4810 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-netns\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512693 4810 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-bin\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512704 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512715 4810 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512726 4810 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-cni-netd\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512802 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512839 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-slash" (OuterVolumeSpecName: "host-slash") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512865 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512888 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512935 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.512962 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.513067 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.520591 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aacbcdc8-7ef5-473b-8055-145f40040d5d-kube-api-access-kx6mf" (OuterVolumeSpecName: "kube-api-access-kx6mf") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "kube-api-access-kx6mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.520814 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.538096 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "aacbcdc8-7ef5-473b-8055-145f40040d5d" (UID: "aacbcdc8-7ef5-473b-8055-145f40040d5d"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.613936 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-node-log\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.613981 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-cni-netd\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614004 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/037365fc-5b0d-4e83-a6e8-eee1e4151047-ovn-node-metrics-cert\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614029 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-run-ovn\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614044 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x8cp\" (UniqueName: \"kubernetes.io/projected/037365fc-5b0d-4e83-a6e8-eee1e4151047-kube-api-access-9x8cp\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614058 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-run-systemd\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614075 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-systemd-units\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614089 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-slash\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614105 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-kubelet\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614120 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614141 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-var-lib-openvswitch\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614161 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-run-netns\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614182 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-etc-openvswitch\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614197 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-run-openvswitch\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/037365fc-5b0d-4e83-a6e8-eee1e4151047-env-overrides\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614228 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-run-ovn-kubernetes\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614250 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/037365fc-5b0d-4e83-a6e8-eee1e4151047-ovnkube-script-lib\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614283 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-cni-bin\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614298 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/037365fc-5b0d-4e83-a6e8-eee1e4151047-ovnkube-config\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614312 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-log-socket\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614350 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx6mf\" (UniqueName: \"kubernetes.io/projected/aacbcdc8-7ef5-473b-8055-145f40040d5d-kube-api-access-kx6mf\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614361 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614369 4810 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aacbcdc8-7ef5-473b-8055-145f40040d5d-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614378 4810 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614386 4810 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-kubelet\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614394 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aacbcdc8-7ef5-473b-8055-145f40040d5d-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614402 4810 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-host-slash\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614410 4810 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614418 4810 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614426 4810 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aacbcdc8-7ef5-473b-8055-145f40040d5d-run-systemd\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614470 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-log-socket\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614511 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-var-lib-openvswitch\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614544 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-run-netns\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614563 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614595 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-etc-openvswitch\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614566 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-run-ovn-kubernetes\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614641 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-run-openvswitch\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614637 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-node-log\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.614672 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-cni-netd\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.615188 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/037365fc-5b0d-4e83-a6e8-eee1e4151047-env-overrides\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.615236 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-cni-bin\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.615320 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/037365fc-5b0d-4e83-a6e8-eee1e4151047-ovnkube-script-lib\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.615365 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-run-systemd\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.615396 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-run-ovn\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.615481 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-slash\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.615540 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-systemd-units\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.615740 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/037365fc-5b0d-4e83-a6e8-eee1e4151047-ovnkube-config\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.615647 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/037365fc-5b0d-4e83-a6e8-eee1e4151047-host-kubelet\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.620150 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/037365fc-5b0d-4e83-a6e8-eee1e4151047-ovn-node-metrics-cert\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.645818 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x8cp\" (UniqueName: \"kubernetes.io/projected/037365fc-5b0d-4e83-a6e8-eee1e4151047-kube-api-access-9x8cp\") pod \"ovnkube-node-5g7w2\" (UID: \"037365fc-5b0d-4e83-a6e8-eee1e4151047\") " pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: I0930 08:12:12.756036 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:12 crc kubenswrapper[4810]: W0930 08:12:12.788866 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod037365fc_5b0d_4e83_a6e8_eee1e4151047.slice/crio-635cb291bccec6b2f772be9c1f5cbbe1d4a3e5b1bad42a9f6534425ca4c1265e WatchSource:0}: Error finding container 635cb291bccec6b2f772be9c1f5cbbe1d4a3e5b1bad42a9f6534425ca4c1265e: Status 404 returned error can't find the container with id 635cb291bccec6b2f772be9c1f5cbbe1d4a3e5b1bad42a9f6534425ca4c1265e Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.357027 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n85f4_f5367714-1633-4695-af4e-4eb6419daa96/kube-multus/2.log" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.359557 4810 generic.go:334] "Generic (PLEG): container finished" podID="037365fc-5b0d-4e83-a6e8-eee1e4151047" containerID="ddf1e21f7db53a2cd3a2393c4e49e17531c52e6be5f0ed73d5e73d4c0f955319" exitCode=0 Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.359694 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" event={"ID":"037365fc-5b0d-4e83-a6e8-eee1e4151047","Type":"ContainerDied","Data":"ddf1e21f7db53a2cd3a2393c4e49e17531c52e6be5f0ed73d5e73d4c0f955319"} Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.359746 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" event={"ID":"037365fc-5b0d-4e83-a6e8-eee1e4151047","Type":"ContainerStarted","Data":"635cb291bccec6b2f772be9c1f5cbbe1d4a3e5b1bad42a9f6534425ca4c1265e"} Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.367100 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovn-acl-logging/0.log" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.367903 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9bkcj_aacbcdc8-7ef5-473b-8055-145f40040d5d/ovn-controller/0.log" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.369216 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" event={"ID":"aacbcdc8-7ef5-473b-8055-145f40040d5d","Type":"ContainerDied","Data":"45d54bd719d7b34f0b30fe78e45e73f656954ceb4b9e69c6dba27b382d9ee125"} Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.369291 4810 scope.go:117] "RemoveContainer" containerID="933fa6a06cdfcff44c3ae380d64b817e01623656491d875120a01cdfbdbdff55" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.369427 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9bkcj" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.398310 4810 scope.go:117] "RemoveContainer" containerID="62ad4bf8bc3b77022c10990405c9b708abac90eedf1e85d0d70a7770944c87b6" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.415958 4810 scope.go:117] "RemoveContainer" containerID="03deffbb1ebefc7af658ede5d185ccedb78d2511ce099f6087ca59c1639eee89" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.433878 4810 scope.go:117] "RemoveContainer" containerID="114d0772e5732da53f5c3cda5322b8afa0e0dd6a75f9d8e6295bfc0a480c809f" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.449850 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9bkcj"] Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.455310 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9bkcj"] Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.458054 4810 scope.go:117] "RemoveContainer" containerID="c6fedea3c51137d842cf157fdda093f947058914b159379f36440fb46bd21f38" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.480984 4810 scope.go:117] "RemoveContainer" containerID="cd8dde6524c3460ac809976838f17562014f0c67e2daf3d90193343dc2da14e5" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.491813 4810 scope.go:117] "RemoveContainer" containerID="5a6708c2a3a2a083ed282d16c8c4f1032f47fa2f216d4c07ce2299530e37aef5" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.540086 4810 scope.go:117] "RemoveContainer" containerID="9e2d4826fe11b797e7ec2f5e5210c2275a03c872c1efb37a38635cff898c1644" Sep 30 08:12:13 crc kubenswrapper[4810]: I0930 08:12:13.553282 4810 scope.go:117] "RemoveContainer" containerID="9cfcf8d7166869987c59350e5d1eb7384989e577b6509bed462f8bb582d08479" Sep 30 08:12:14 crc kubenswrapper[4810]: I0930 08:12:14.377572 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" event={"ID":"037365fc-5b0d-4e83-a6e8-eee1e4151047","Type":"ContainerStarted","Data":"1f3feacea923d9c9ff92e40a5dae3a43977153bfe045dc139a597bc3f75e2f75"} Sep 30 08:12:14 crc kubenswrapper[4810]: I0930 08:12:14.378699 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" event={"ID":"037365fc-5b0d-4e83-a6e8-eee1e4151047","Type":"ContainerStarted","Data":"29ef66f9a1ef00f7c14dd74de68cb1508fbbf1ac425edebebed5cf1594788892"} Sep 30 08:12:14 crc kubenswrapper[4810]: I0930 08:12:14.378810 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" event={"ID":"037365fc-5b0d-4e83-a6e8-eee1e4151047","Type":"ContainerStarted","Data":"66ec51e0b54060a62f1b829e1eb099fc0a6ccc06f039a4bf04673815677a35f1"} Sep 30 08:12:14 crc kubenswrapper[4810]: I0930 08:12:14.378827 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" event={"ID":"037365fc-5b0d-4e83-a6e8-eee1e4151047","Type":"ContainerStarted","Data":"90f806f75d4dd01c4cd6ffbbf9a8a9ebe13cfedf0fc8c8ca56cd06e063c37deb"} Sep 30 08:12:14 crc kubenswrapper[4810]: I0930 08:12:14.378844 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" event={"ID":"037365fc-5b0d-4e83-a6e8-eee1e4151047","Type":"ContainerStarted","Data":"f6206ea1a23b0b4c2696f8e99461072ce74e4e577f6dc3288333168042483fb0"} Sep 30 08:12:14 crc kubenswrapper[4810]: I0930 08:12:14.378859 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" event={"ID":"037365fc-5b0d-4e83-a6e8-eee1e4151047","Type":"ContainerStarted","Data":"4cb9def25a4ca43e09009d5e23b78e14e7cdc30e799b43b8c7c5cd8e658dfa00"} Sep 30 08:12:15 crc kubenswrapper[4810]: I0930 08:12:15.320574 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aacbcdc8-7ef5-473b-8055-145f40040d5d" path="/var/lib/kubelet/pods/aacbcdc8-7ef5-473b-8055-145f40040d5d/volumes" Sep 30 08:12:16 crc kubenswrapper[4810]: I0930 08:12:16.399334 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" event={"ID":"037365fc-5b0d-4e83-a6e8-eee1e4151047","Type":"ContainerStarted","Data":"8e163c09e77ce259b3980e3783ffe8be33b20ede0bc8683d071431aa533742de"} Sep 30 08:12:19 crc kubenswrapper[4810]: I0930 08:12:19.427730 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" event={"ID":"037365fc-5b0d-4e83-a6e8-eee1e4151047","Type":"ContainerStarted","Data":"b51716291634be6be2d156e6fe85fe3513bd39891307e835d2d0c1506235cd4d"} Sep 30 08:12:19 crc kubenswrapper[4810]: I0930 08:12:19.428743 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:19 crc kubenswrapper[4810]: I0930 08:12:19.428775 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:19 crc kubenswrapper[4810]: I0930 08:12:19.480878 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" podStartSLOduration=7.480854781 podStartE2EDuration="7.480854781s" podCreationTimestamp="2025-09-30 08:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:12:19.474401977 +0000 UTC m=+562.926601244" watchObservedRunningTime="2025-09-30 08:12:19.480854781 +0000 UTC m=+562.933054048" Sep 30 08:12:19 crc kubenswrapper[4810]: I0930 08:12:19.487418 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:20 crc kubenswrapper[4810]: I0930 08:12:20.436472 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:20 crc kubenswrapper[4810]: I0930 08:12:20.522580 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:26 crc kubenswrapper[4810]: I0930 08:12:26.306565 4810 scope.go:117] "RemoveContainer" containerID="8a0c05c5c70b22d510babd5c578400d035ba077b50083c564906b0e8dc68be45" Sep 30 08:12:26 crc kubenswrapper[4810]: E0930 08:12:26.307467 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-n85f4_openshift-multus(f5367714-1633-4695-af4e-4eb6419daa96)\"" pod="openshift-multus/multus-n85f4" podUID="f5367714-1633-4695-af4e-4eb6419daa96" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.593146 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs"] Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.595132 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.600919 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.607684 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs"] Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.692952 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9j2h\" (UniqueName: \"kubernetes.io/projected/9a4af0f1-83d1-4384-a16b-cb281e582ecb-kube-api-access-q9j2h\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.693062 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.693087 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.794159 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9j2h\" (UniqueName: \"kubernetes.io/projected/9a4af0f1-83d1-4384-a16b-cb281e582ecb-kube-api-access-q9j2h\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.794213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.794234 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.794705 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.794939 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.816509 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9j2h\" (UniqueName: \"kubernetes.io/projected/9a4af0f1-83d1-4384-a16b-cb281e582ecb-kube-api-access-q9j2h\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: I0930 08:12:39.924975 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: E0930 08:12:39.967571 4810 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace_9a4af0f1-83d1-4384-a16b-cb281e582ecb_0(92a704a9be7aaec2f6f9a0ac361d3ba4076620b8ca2e33a291b3307ea57e77ce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 08:12:39 crc kubenswrapper[4810]: E0930 08:12:39.967655 4810 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace_9a4af0f1-83d1-4384-a16b-cb281e582ecb_0(92a704a9be7aaec2f6f9a0ac361d3ba4076620b8ca2e33a291b3307ea57e77ce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: E0930 08:12:39.967680 4810 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace_9a4af0f1-83d1-4384-a16b-cb281e582ecb_0(92a704a9be7aaec2f6f9a0ac361d3ba4076620b8ca2e33a291b3307ea57e77ce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:39 crc kubenswrapper[4810]: E0930 08:12:39.967732 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace(9a4af0f1-83d1-4384-a16b-cb281e582ecb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace(9a4af0f1-83d1-4384-a16b-cb281e582ecb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace_9a4af0f1-83d1-4384-a16b-cb281e582ecb_0(92a704a9be7aaec2f6f9a0ac361d3ba4076620b8ca2e33a291b3307ea57e77ce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" podUID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" Sep 30 08:12:40 crc kubenswrapper[4810]: I0930 08:12:40.574364 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:40 crc kubenswrapper[4810]: I0930 08:12:40.575005 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:40 crc kubenswrapper[4810]: E0930 08:12:40.597458 4810 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace_9a4af0f1-83d1-4384-a16b-cb281e582ecb_0(8d0d708798f929b276682d7177602de008fef7baeaedfa88664069851c06ccfd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 08:12:40 crc kubenswrapper[4810]: E0930 08:12:40.597751 4810 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace_9a4af0f1-83d1-4384-a16b-cb281e582ecb_0(8d0d708798f929b276682d7177602de008fef7baeaedfa88664069851c06ccfd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:40 crc kubenswrapper[4810]: E0930 08:12:40.597852 4810 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace_9a4af0f1-83d1-4384-a16b-cb281e582ecb_0(8d0d708798f929b276682d7177602de008fef7baeaedfa88664069851c06ccfd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:40 crc kubenswrapper[4810]: E0930 08:12:40.598052 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace(9a4af0f1-83d1-4384-a16b-cb281e582ecb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace(9a4af0f1-83d1-4384-a16b-cb281e582ecb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_openshift-marketplace_9a4af0f1-83d1-4384-a16b-cb281e582ecb_0(8d0d708798f929b276682d7177602de008fef7baeaedfa88664069851c06ccfd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" podUID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" Sep 30 08:12:41 crc kubenswrapper[4810]: I0930 08:12:41.306714 4810 scope.go:117] "RemoveContainer" containerID="8a0c05c5c70b22d510babd5c578400d035ba077b50083c564906b0e8dc68be45" Sep 30 08:12:41 crc kubenswrapper[4810]: I0930 08:12:41.583575 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n85f4_f5367714-1633-4695-af4e-4eb6419daa96/kube-multus/2.log" Sep 30 08:12:41 crc kubenswrapper[4810]: I0930 08:12:41.584416 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n85f4" event={"ID":"f5367714-1633-4695-af4e-4eb6419daa96","Type":"ContainerStarted","Data":"f89dff23b5ae2172de01b80a96e751e2945af690ec39d4d5fa0689fa09d70e7d"} Sep 30 08:12:42 crc kubenswrapper[4810]: I0930 08:12:42.791939 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5g7w2" Sep 30 08:12:45 crc kubenswrapper[4810]: I0930 08:12:45.912154 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:12:45 crc kubenswrapper[4810]: I0930 08:12:45.912581 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:12:52 crc kubenswrapper[4810]: I0930 08:12:52.306428 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:52 crc kubenswrapper[4810]: I0930 08:12:52.307957 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:52 crc kubenswrapper[4810]: I0930 08:12:52.568102 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs"] Sep 30 08:12:52 crc kubenswrapper[4810]: I0930 08:12:52.665521 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" event={"ID":"9a4af0f1-83d1-4384-a16b-cb281e582ecb","Type":"ContainerStarted","Data":"14e8f13362e819d48bac3fb63b00657a1f96890b6675328fe5736de4d8147916"} Sep 30 08:12:53 crc kubenswrapper[4810]: I0930 08:12:53.676582 4810 generic.go:334] "Generic (PLEG): container finished" podID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" containerID="d77c36b58963280b136fa8c4ad405a63e3547a2ab9ec5036847a21c0542a9057" exitCode=0 Sep 30 08:12:53 crc kubenswrapper[4810]: I0930 08:12:53.676684 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" event={"ID":"9a4af0f1-83d1-4384-a16b-cb281e582ecb","Type":"ContainerDied","Data":"d77c36b58963280b136fa8c4ad405a63e3547a2ab9ec5036847a21c0542a9057"} Sep 30 08:12:55 crc kubenswrapper[4810]: I0930 08:12:55.693743 4810 generic.go:334] "Generic (PLEG): container finished" podID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" containerID="8f240ed4d7beb6a9e7e85d03384a2d8eca6dab4a6ca912d0d213b248a426501e" exitCode=0 Sep 30 08:12:55 crc kubenswrapper[4810]: I0930 08:12:55.693841 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" event={"ID":"9a4af0f1-83d1-4384-a16b-cb281e582ecb","Type":"ContainerDied","Data":"8f240ed4d7beb6a9e7e85d03384a2d8eca6dab4a6ca912d0d213b248a426501e"} Sep 30 08:12:56 crc kubenswrapper[4810]: I0930 08:12:56.710584 4810 generic.go:334] "Generic (PLEG): container finished" podID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" containerID="c14d4bb1bd5dfc6f3414004833696256cf021b2695910aecc133088a9e1a639f" exitCode=0 Sep 30 08:12:56 crc kubenswrapper[4810]: I0930 08:12:56.710726 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" event={"ID":"9a4af0f1-83d1-4384-a16b-cb281e582ecb","Type":"ContainerDied","Data":"c14d4bb1bd5dfc6f3414004833696256cf021b2695910aecc133088a9e1a639f"} Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.041592 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.145519 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-bundle\") pod \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.145766 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9j2h\" (UniqueName: \"kubernetes.io/projected/9a4af0f1-83d1-4384-a16b-cb281e582ecb-kube-api-access-q9j2h\") pod \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.146013 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-util\") pod \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\" (UID: \"9a4af0f1-83d1-4384-a16b-cb281e582ecb\") " Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.148677 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-bundle" (OuterVolumeSpecName: "bundle") pod "9a4af0f1-83d1-4384-a16b-cb281e582ecb" (UID: "9a4af0f1-83d1-4384-a16b-cb281e582ecb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.155467 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a4af0f1-83d1-4384-a16b-cb281e582ecb-kube-api-access-q9j2h" (OuterVolumeSpecName: "kube-api-access-q9j2h") pod "9a4af0f1-83d1-4384-a16b-cb281e582ecb" (UID: "9a4af0f1-83d1-4384-a16b-cb281e582ecb"). InnerVolumeSpecName "kube-api-access-q9j2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.247947 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.249107 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9j2h\" (UniqueName: \"kubernetes.io/projected/9a4af0f1-83d1-4384-a16b-cb281e582ecb-kube-api-access-q9j2h\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.251253 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-util" (OuterVolumeSpecName: "util") pod "9a4af0f1-83d1-4384-a16b-cb281e582ecb" (UID: "9a4af0f1-83d1-4384-a16b-cb281e582ecb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.351577 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a4af0f1-83d1-4384-a16b-cb281e582ecb-util\") on node \"crc\" DevicePath \"\"" Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.729609 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" event={"ID":"9a4af0f1-83d1-4384-a16b-cb281e582ecb","Type":"ContainerDied","Data":"14e8f13362e819d48bac3fb63b00657a1f96890b6675328fe5736de4d8147916"} Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.729668 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14e8f13362e819d48bac3fb63b00657a1f96890b6675328fe5736de4d8147916" Sep 30 08:12:58 crc kubenswrapper[4810]: I0930 08:12:58.729717 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.565055 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-9cddq"] Sep 30 08:13:06 crc kubenswrapper[4810]: E0930 08:13:06.565834 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" containerName="extract" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.565851 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" containerName="extract" Sep 30 08:13:06 crc kubenswrapper[4810]: E0930 08:13:06.565874 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" containerName="pull" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.565880 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" containerName="pull" Sep 30 08:13:06 crc kubenswrapper[4810]: E0930 08:13:06.565890 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" containerName="util" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.565897 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" containerName="util" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.566011 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4af0f1-83d1-4384-a16b-cb281e582ecb" containerName="extract" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.566444 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cddq" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.569923 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.569978 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-nkkvq" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.570120 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.580234 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-9cddq"] Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.646668 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8"] Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.647731 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.652389 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.652639 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-lhqqt" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.654961 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p"] Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.655627 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.663117 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8"] Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.665938 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhnj8\" (UniqueName: \"kubernetes.io/projected/360a4c93-d93b-4caf-b11d-a2185a5179d3-kube-api-access-hhnj8\") pod \"obo-prometheus-operator-7c8cf85677-9cddq\" (UID: \"360a4c93-d93b-4caf-b11d-a2185a5179d3\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cddq" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.678119 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p"] Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.767657 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhnj8\" (UniqueName: \"kubernetes.io/projected/360a4c93-d93b-4caf-b11d-a2185a5179d3-kube-api-access-hhnj8\") pod \"obo-prometheus-operator-7c8cf85677-9cddq\" (UID: \"360a4c93-d93b-4caf-b11d-a2185a5179d3\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cddq" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.767739 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e41015ca-c417-4076-9895-abff0488bfd7-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p\" (UID: \"e41015ca-c417-4076-9895-abff0488bfd7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.767783 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ea01519-d717-45ef-be26-b429eb2c6cdb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8\" (UID: \"2ea01519-d717-45ef-be26-b429eb2c6cdb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.767816 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ea01519-d717-45ef-be26-b429eb2c6cdb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8\" (UID: \"2ea01519-d717-45ef-be26-b429eb2c6cdb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.767906 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e41015ca-c417-4076-9895-abff0488bfd7-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p\" (UID: \"e41015ca-c417-4076-9895-abff0488bfd7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.785530 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-nq2k6"] Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.786365 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.786961 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhnj8\" (UniqueName: \"kubernetes.io/projected/360a4c93-d93b-4caf-b11d-a2185a5179d3-kube-api-access-hhnj8\") pod \"obo-prometheus-operator-7c8cf85677-9cddq\" (UID: \"360a4c93-d93b-4caf-b11d-a2185a5179d3\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cddq" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.788786 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-bn9vk" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.789038 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.809295 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-nq2k6"] Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.868989 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e41015ca-c417-4076-9895-abff0488bfd7-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p\" (UID: \"e41015ca-c417-4076-9895-abff0488bfd7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.869040 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0da3209d-75c8-4eda-97f4-b6cfeac48f62-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-nq2k6\" (UID: \"0da3209d-75c8-4eda-97f4-b6cfeac48f62\") " pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.869085 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e41015ca-c417-4076-9895-abff0488bfd7-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p\" (UID: \"e41015ca-c417-4076-9895-abff0488bfd7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.869112 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ea01519-d717-45ef-be26-b429eb2c6cdb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8\" (UID: \"2ea01519-d717-45ef-be26-b429eb2c6cdb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.869135 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ea01519-d717-45ef-be26-b429eb2c6cdb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8\" (UID: \"2ea01519-d717-45ef-be26-b429eb2c6cdb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.869190 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk2z7\" (UniqueName: \"kubernetes.io/projected/0da3209d-75c8-4eda-97f4-b6cfeac48f62-kube-api-access-fk2z7\") pod \"observability-operator-cc5f78dfc-nq2k6\" (UID: \"0da3209d-75c8-4eda-97f4-b6cfeac48f62\") " pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.872917 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e41015ca-c417-4076-9895-abff0488bfd7-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p\" (UID: \"e41015ca-c417-4076-9895-abff0488bfd7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.872919 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ea01519-d717-45ef-be26-b429eb2c6cdb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8\" (UID: \"2ea01519-d717-45ef-be26-b429eb2c6cdb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.873235 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e41015ca-c417-4076-9895-abff0488bfd7-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p\" (UID: \"e41015ca-c417-4076-9895-abff0488bfd7\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.873546 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ea01519-d717-45ef-be26-b429eb2c6cdb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8\" (UID: \"2ea01519-d717-45ef-be26-b429eb2c6cdb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.910131 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cddq" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.932651 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-kj7xb"] Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.933422 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.935866 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-l2qzl" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.944286 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-kj7xb"] Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.969484 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.969920 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk2z7\" (UniqueName: \"kubernetes.io/projected/0da3209d-75c8-4eda-97f4-b6cfeac48f62-kube-api-access-fk2z7\") pod \"observability-operator-cc5f78dfc-nq2k6\" (UID: \"0da3209d-75c8-4eda-97f4-b6cfeac48f62\") " pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.969975 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0da3209d-75c8-4eda-97f4-b6cfeac48f62-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-nq2k6\" (UID: \"0da3209d-75c8-4eda-97f4-b6cfeac48f62\") " pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.972745 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0da3209d-75c8-4eda-97f4-b6cfeac48f62-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-nq2k6\" (UID: \"0da3209d-75c8-4eda-97f4-b6cfeac48f62\") " pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.979278 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p" Sep 30 08:13:06 crc kubenswrapper[4810]: I0930 08:13:06.985707 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk2z7\" (UniqueName: \"kubernetes.io/projected/0da3209d-75c8-4eda-97f4-b6cfeac48f62-kube-api-access-fk2z7\") pod \"observability-operator-cc5f78dfc-nq2k6\" (UID: \"0da3209d-75c8-4eda-97f4-b6cfeac48f62\") " pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.071966 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xbxv\" (UniqueName: \"kubernetes.io/projected/5f51c1dc-ac89-4a84-8897-d2e9770baec9-kube-api-access-9xbxv\") pod \"perses-operator-54bc95c9fb-kj7xb\" (UID: \"5f51c1dc-ac89-4a84-8897-d2e9770baec9\") " pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.072331 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f51c1dc-ac89-4a84-8897-d2e9770baec9-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-kj7xb\" (UID: \"5f51c1dc-ac89-4a84-8897-d2e9770baec9\") " pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.121828 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.168176 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-9cddq"] Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.172935 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f51c1dc-ac89-4a84-8897-d2e9770baec9-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-kj7xb\" (UID: \"5f51c1dc-ac89-4a84-8897-d2e9770baec9\") " pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.172983 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xbxv\" (UniqueName: \"kubernetes.io/projected/5f51c1dc-ac89-4a84-8897-d2e9770baec9-kube-api-access-9xbxv\") pod \"perses-operator-54bc95c9fb-kj7xb\" (UID: \"5f51c1dc-ac89-4a84-8897-d2e9770baec9\") " pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.173845 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f51c1dc-ac89-4a84-8897-d2e9770baec9-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-kj7xb\" (UID: \"5f51c1dc-ac89-4a84-8897-d2e9770baec9\") " pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" Sep 30 08:13:07 crc kubenswrapper[4810]: W0930 08:13:07.186833 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod360a4c93_d93b_4caf_b11d_a2185a5179d3.slice/crio-648a39031832e963df8d232ee5016d342a668755d810eb4b7a5746573b43ba8c WatchSource:0}: Error finding container 648a39031832e963df8d232ee5016d342a668755d810eb4b7a5746573b43ba8c: Status 404 returned error can't find the container with id 648a39031832e963df8d232ee5016d342a668755d810eb4b7a5746573b43ba8c Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.190312 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xbxv\" (UniqueName: \"kubernetes.io/projected/5f51c1dc-ac89-4a84-8897-d2e9770baec9-kube-api-access-9xbxv\") pod \"perses-operator-54bc95c9fb-kj7xb\" (UID: \"5f51c1dc-ac89-4a84-8897-d2e9770baec9\") " pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.222980 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8"] Sep 30 08:13:07 crc kubenswrapper[4810]: W0930 08:13:07.243646 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ea01519_d717_45ef_be26_b429eb2c6cdb.slice/crio-c5cafeacc26771bcc0c9830f5828cd9a40e46d849ea8892614a78e26c66db629 WatchSource:0}: Error finding container c5cafeacc26771bcc0c9830f5828cd9a40e46d849ea8892614a78e26c66db629: Status 404 returned error can't find the container with id c5cafeacc26771bcc0c9830f5828cd9a40e46d849ea8892614a78e26c66db629 Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.270785 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.286241 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p"] Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.512809 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-kj7xb"] Sep 30 08:13:07 crc kubenswrapper[4810]: W0930 08:13:07.520073 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f51c1dc_ac89_4a84_8897_d2e9770baec9.slice/crio-092d7b442aa0c5bf28dd32bfe2cf601bb5a7cd1e2e03b82393a6c6a0e7a6231e WatchSource:0}: Error finding container 092d7b442aa0c5bf28dd32bfe2cf601bb5a7cd1e2e03b82393a6c6a0e7a6231e: Status 404 returned error can't find the container with id 092d7b442aa0c5bf28dd32bfe2cf601bb5a7cd1e2e03b82393a6c6a0e7a6231e Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.589115 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-nq2k6"] Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.801254 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8" event={"ID":"2ea01519-d717-45ef-be26-b429eb2c6cdb","Type":"ContainerStarted","Data":"c5cafeacc26771bcc0c9830f5828cd9a40e46d849ea8892614a78e26c66db629"} Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.802387 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cddq" event={"ID":"360a4c93-d93b-4caf-b11d-a2185a5179d3","Type":"ContainerStarted","Data":"648a39031832e963df8d232ee5016d342a668755d810eb4b7a5746573b43ba8c"} Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.803871 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" event={"ID":"5f51c1dc-ac89-4a84-8897-d2e9770baec9","Type":"ContainerStarted","Data":"092d7b442aa0c5bf28dd32bfe2cf601bb5a7cd1e2e03b82393a6c6a0e7a6231e"} Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.804890 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" event={"ID":"0da3209d-75c8-4eda-97f4-b6cfeac48f62","Type":"ContainerStarted","Data":"fd8d0ad2093621f30e9bb94ab7768b44e7eedc97e30999a7276dd64b52692a32"} Sep 30 08:13:07 crc kubenswrapper[4810]: I0930 08:13:07.805926 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p" event={"ID":"e41015ca-c417-4076-9895-abff0488bfd7","Type":"ContainerStarted","Data":"8e37acb56a31bdee9d4ea981e2030bdbd6facfe74b081bf6d5d4067a79b0cdf8"} Sep 30 08:13:15 crc kubenswrapper[4810]: I0930 08:13:15.911805 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:13:15 crc kubenswrapper[4810]: I0930 08:13:15.912053 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:13:23 crc kubenswrapper[4810]: E0930 08:13:23.506574 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e" Sep 30 08:13:23 crc kubenswrapper[4810]: E0930 08:13:23.507417 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e,Command:[],Args:[--namespace=$(NAMESPACE) --images=alertmanager=$(RELATED_IMAGE_ALERTMANAGER) --images=prometheus=$(RELATED_IMAGE_PROMETHEUS) --images=thanos=$(RELATED_IMAGE_THANOS) --images=perses=$(RELATED_IMAGE_PERSES) --images=ui-dashboards=$(RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN) --images=ui-distributed-tracing=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN) --images=ui-distributed-tracing-pf5=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5) --images=ui-distributed-tracing-pf4=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4) --images=ui-logging=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN) --images=ui-logging-pf4=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4) --images=ui-troubleshooting-panel=$(RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN) --images=ui-monitoring=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN) --images=ui-monitoring-pf5=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5) --images=korrel8r=$(RELATED_IMAGE_KORREL8R) --images=health-analyzer=$(RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER) --openshift.enabled=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER,Value:registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS,Value:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_THANOS,Value:registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PERSES,Value:registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KORREL8R,Value:registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER,Value:registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{400 -3} {} 400m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:observability-operator-tls,ReadOnly:true,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fk2z7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod observability-operator-cc5f78dfc-nq2k6_openshift-operators(0da3209d-75c8-4eda-97f4-b6cfeac48f62): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 08:13:23 crc kubenswrapper[4810]: E0930 08:13:23.508701 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" podUID="0da3209d-75c8-4eda-97f4-b6cfeac48f62" Sep 30 08:13:23 crc kubenswrapper[4810]: E0930 08:13:23.967069 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e\\\"\"" pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" podUID="0da3209d-75c8-4eda-97f4-b6cfeac48f62" Sep 30 08:13:24 crc kubenswrapper[4810]: E0930 08:13:24.127185 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c" Sep 30 08:13:24 crc kubenswrapper[4810]: E0930 08:13:24.127478 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9xbxv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-54bc95c9fb-kj7xb_openshift-operators(5f51c1dc-ac89-4a84-8897-d2e9770baec9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 08:13:24 crc kubenswrapper[4810]: E0930 08:13:24.128716 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" podUID="5f51c1dc-ac89-4a84-8897-d2e9770baec9" Sep 30 08:13:24 crc kubenswrapper[4810]: I0930 08:13:24.970401 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8" event={"ID":"2ea01519-d717-45ef-be26-b429eb2c6cdb","Type":"ContainerStarted","Data":"a821604115b41d8d01ebf70c4bf6d9cb4a6b023772c6ddf8dd239ff97634d448"} Sep 30 08:13:24 crc kubenswrapper[4810]: I0930 08:13:24.973324 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cddq" event={"ID":"360a4c93-d93b-4caf-b11d-a2185a5179d3","Type":"ContainerStarted","Data":"22f6bd441e5ee47597a218ad2e19d456fedbc63e17c61a4553dc252b061e6424"} Sep 30 08:13:24 crc kubenswrapper[4810]: I0930 08:13:24.976149 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p" event={"ID":"e41015ca-c417-4076-9895-abff0488bfd7","Type":"ContainerStarted","Data":"4b6c0d47a81f730f512a7b0b85b7493354b700a16680f8adbc3e49d9054d7a64"} Sep 30 08:13:24 crc kubenswrapper[4810]: E0930 08:13:24.976544 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c\\\"\"" pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" podUID="5f51c1dc-ac89-4a84-8897-d2e9770baec9" Sep 30 08:13:24 crc kubenswrapper[4810]: I0930 08:13:24.994654 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8" podStartSLOduration=2.12185048 podStartE2EDuration="18.994639657s" podCreationTimestamp="2025-09-30 08:13:06 +0000 UTC" firstStartedPulling="2025-09-30 08:13:07.255932093 +0000 UTC m=+610.708131350" lastFinishedPulling="2025-09-30 08:13:24.12872126 +0000 UTC m=+627.580920527" observedRunningTime="2025-09-30 08:13:24.992755823 +0000 UTC m=+628.444955090" watchObservedRunningTime="2025-09-30 08:13:24.994639657 +0000 UTC m=+628.446838924" Sep 30 08:13:25 crc kubenswrapper[4810]: I0930 08:13:25.032070 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cddq" podStartSLOduration=2.083361296 podStartE2EDuration="19.03205277s" podCreationTimestamp="2025-09-30 08:13:06 +0000 UTC" firstStartedPulling="2025-09-30 08:13:07.192443112 +0000 UTC m=+610.644642379" lastFinishedPulling="2025-09-30 08:13:24.141134556 +0000 UTC m=+627.593333853" observedRunningTime="2025-09-30 08:13:25.026403468 +0000 UTC m=+628.478602735" watchObservedRunningTime="2025-09-30 08:13:25.03205277 +0000 UTC m=+628.484252037" Sep 30 08:13:35 crc kubenswrapper[4810]: I0930 08:13:35.333029 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p" podStartSLOduration=12.492879799 podStartE2EDuration="29.333000058s" podCreationTimestamp="2025-09-30 08:13:06 +0000 UTC" firstStartedPulling="2025-09-30 08:13:07.312128706 +0000 UTC m=+610.764327973" lastFinishedPulling="2025-09-30 08:13:24.152248925 +0000 UTC m=+627.604448232" observedRunningTime="2025-09-30 08:13:25.052160756 +0000 UTC m=+628.504360013" watchObservedRunningTime="2025-09-30 08:13:35.333000058 +0000 UTC m=+638.785199365" Sep 30 08:13:37 crc kubenswrapper[4810]: I0930 08:13:37.059893 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" event={"ID":"5f51c1dc-ac89-4a84-8897-d2e9770baec9","Type":"ContainerStarted","Data":"25b16bfd45c6b258a454defaf991af222910e5daab4b135f9c9d90ad0c01ff29"} Sep 30 08:13:37 crc kubenswrapper[4810]: I0930 08:13:37.060580 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" Sep 30 08:13:37 crc kubenswrapper[4810]: I0930 08:13:37.068809 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" event={"ID":"0da3209d-75c8-4eda-97f4-b6cfeac48f62","Type":"ContainerStarted","Data":"bc19297b1247dffbc2628953401a79456b564cc69760db364c737a6fa3e0f336"} Sep 30 08:13:37 crc kubenswrapper[4810]: I0930 08:13:37.070244 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" Sep 30 08:13:37 crc kubenswrapper[4810]: I0930 08:13:37.086942 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" Sep 30 08:13:37 crc kubenswrapper[4810]: I0930 08:13:37.096306 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" podStartSLOduration=2.568521908 podStartE2EDuration="31.096262841s" podCreationTimestamp="2025-09-30 08:13:06 +0000 UTC" firstStartedPulling="2025-09-30 08:13:07.522670641 +0000 UTC m=+610.974869918" lastFinishedPulling="2025-09-30 08:13:36.050411544 +0000 UTC m=+639.502610851" observedRunningTime="2025-09-30 08:13:37.088289233 +0000 UTC m=+640.540488520" watchObservedRunningTime="2025-09-30 08:13:37.096262841 +0000 UTC m=+640.548462118" Sep 30 08:13:37 crc kubenswrapper[4810]: I0930 08:13:37.126980 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-nq2k6" podStartSLOduration=2.430091455 podStartE2EDuration="31.126959922s" podCreationTimestamp="2025-09-30 08:13:06 +0000 UTC" firstStartedPulling="2025-09-30 08:13:07.591462851 +0000 UTC m=+611.043662118" lastFinishedPulling="2025-09-30 08:13:36.288331308 +0000 UTC m=+639.740530585" observedRunningTime="2025-09-30 08:13:37.126180639 +0000 UTC m=+640.578379986" watchObservedRunningTime="2025-09-30 08:13:37.126959922 +0000 UTC m=+640.579159199" Sep 30 08:13:45 crc kubenswrapper[4810]: I0930 08:13:45.911704 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:13:45 crc kubenswrapper[4810]: I0930 08:13:45.912523 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:13:45 crc kubenswrapper[4810]: I0930 08:13:45.912617 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:13:45 crc kubenswrapper[4810]: I0930 08:13:45.913763 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e8713aca2fba3818ce4206358c20335307a8230c5f24759e0b223332c10c94d"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:13:45 crc kubenswrapper[4810]: I0930 08:13:45.913891 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://3e8713aca2fba3818ce4206358c20335307a8230c5f24759e0b223332c10c94d" gracePeriod=600 Sep 30 08:13:46 crc kubenswrapper[4810]: I0930 08:13:46.134083 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="3e8713aca2fba3818ce4206358c20335307a8230c5f24759e0b223332c10c94d" exitCode=0 Sep 30 08:13:46 crc kubenswrapper[4810]: I0930 08:13:46.134181 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"3e8713aca2fba3818ce4206358c20335307a8230c5f24759e0b223332c10c94d"} Sep 30 08:13:46 crc kubenswrapper[4810]: I0930 08:13:46.134444 4810 scope.go:117] "RemoveContainer" containerID="33485b79f13ae3509d05af432da899150f9b5186a42141758f7a50e6033f5705" Sep 30 08:13:47 crc kubenswrapper[4810]: I0930 08:13:47.146678 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"b2b0b0f4ca79c884d37aac25744a2d16466a6c0fbe29d05bbf5e9792b146f420"} Sep 30 08:13:47 crc kubenswrapper[4810]: I0930 08:13:47.275087 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-kj7xb" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.338615 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9"] Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.340297 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.344431 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.350439 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9"] Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.386576 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jshm\" (UniqueName: \"kubernetes.io/projected/df44374d-4490-461f-aeed-05fae8096705-kube-api-access-8jshm\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.386639 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.386746 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.487790 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jshm\" (UniqueName: \"kubernetes.io/projected/df44374d-4490-461f-aeed-05fae8096705-kube-api-access-8jshm\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.487866 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.487898 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.488468 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.488610 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.516869 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jshm\" (UniqueName: \"kubernetes.io/projected/df44374d-4490-461f-aeed-05fae8096705-kube-api-access-8jshm\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.657479 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:05 crc kubenswrapper[4810]: I0930 08:14:05.932174 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9"] Sep 30 08:14:06 crc kubenswrapper[4810]: I0930 08:14:06.318715 4810 generic.go:334] "Generic (PLEG): container finished" podID="df44374d-4490-461f-aeed-05fae8096705" containerID="e465a08d802d42d978e53987b788dfc8fb863e4f770b9636076909589653cd84" exitCode=0 Sep 30 08:14:06 crc kubenswrapper[4810]: I0930 08:14:06.318825 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" event={"ID":"df44374d-4490-461f-aeed-05fae8096705","Type":"ContainerDied","Data":"e465a08d802d42d978e53987b788dfc8fb863e4f770b9636076909589653cd84"} Sep 30 08:14:06 crc kubenswrapper[4810]: I0930 08:14:06.319540 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" event={"ID":"df44374d-4490-461f-aeed-05fae8096705","Type":"ContainerStarted","Data":"27636f2422460fa31d9d450f77dc80075b12edc005b84fc0a7e19de235732624"} Sep 30 08:14:08 crc kubenswrapper[4810]: I0930 08:14:08.341779 4810 generic.go:334] "Generic (PLEG): container finished" podID="df44374d-4490-461f-aeed-05fae8096705" containerID="ed3c42abe09a4eac1fb0f7e732dc804b47de466847ebebe2190c8a5a2c86f7b0" exitCode=0 Sep 30 08:14:08 crc kubenswrapper[4810]: I0930 08:14:08.341878 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" event={"ID":"df44374d-4490-461f-aeed-05fae8096705","Type":"ContainerDied","Data":"ed3c42abe09a4eac1fb0f7e732dc804b47de466847ebebe2190c8a5a2c86f7b0"} Sep 30 08:14:09 crc kubenswrapper[4810]: I0930 08:14:09.352542 4810 generic.go:334] "Generic (PLEG): container finished" podID="df44374d-4490-461f-aeed-05fae8096705" containerID="4178a5aa7f0e78145d6436338c1210721d142327e3867b9ac2ace99cee2872b1" exitCode=0 Sep 30 08:14:09 crc kubenswrapper[4810]: I0930 08:14:09.352604 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" event={"ID":"df44374d-4490-461f-aeed-05fae8096705","Type":"ContainerDied","Data":"4178a5aa7f0e78145d6436338c1210721d142327e3867b9ac2ace99cee2872b1"} Sep 30 08:14:10 crc kubenswrapper[4810]: I0930 08:14:10.622984 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:10 crc kubenswrapper[4810]: I0930 08:14:10.793662 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-util\") pod \"df44374d-4490-461f-aeed-05fae8096705\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " Sep 30 08:14:10 crc kubenswrapper[4810]: I0930 08:14:10.793792 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-bundle\") pod \"df44374d-4490-461f-aeed-05fae8096705\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " Sep 30 08:14:10 crc kubenswrapper[4810]: I0930 08:14:10.793872 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jshm\" (UniqueName: \"kubernetes.io/projected/df44374d-4490-461f-aeed-05fae8096705-kube-api-access-8jshm\") pod \"df44374d-4490-461f-aeed-05fae8096705\" (UID: \"df44374d-4490-461f-aeed-05fae8096705\") " Sep 30 08:14:10 crc kubenswrapper[4810]: I0930 08:14:10.794704 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-bundle" (OuterVolumeSpecName: "bundle") pod "df44374d-4490-461f-aeed-05fae8096705" (UID: "df44374d-4490-461f-aeed-05fae8096705"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:14:10 crc kubenswrapper[4810]: I0930 08:14:10.803136 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df44374d-4490-461f-aeed-05fae8096705-kube-api-access-8jshm" (OuterVolumeSpecName: "kube-api-access-8jshm") pod "df44374d-4490-461f-aeed-05fae8096705" (UID: "df44374d-4490-461f-aeed-05fae8096705"). InnerVolumeSpecName "kube-api-access-8jshm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:14:10 crc kubenswrapper[4810]: I0930 08:14:10.806402 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-util" (OuterVolumeSpecName: "util") pod "df44374d-4490-461f-aeed-05fae8096705" (UID: "df44374d-4490-461f-aeed-05fae8096705"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:14:10 crc kubenswrapper[4810]: I0930 08:14:10.895312 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-util\") on node \"crc\" DevicePath \"\"" Sep 30 08:14:10 crc kubenswrapper[4810]: I0930 08:14:10.895344 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df44374d-4490-461f-aeed-05fae8096705-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:14:10 crc kubenswrapper[4810]: I0930 08:14:10.895353 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jshm\" (UniqueName: \"kubernetes.io/projected/df44374d-4490-461f-aeed-05fae8096705-kube-api-access-8jshm\") on node \"crc\" DevicePath \"\"" Sep 30 08:14:11 crc kubenswrapper[4810]: I0930 08:14:11.371231 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" event={"ID":"df44374d-4490-461f-aeed-05fae8096705","Type":"ContainerDied","Data":"27636f2422460fa31d9d450f77dc80075b12edc005b84fc0a7e19de235732624"} Sep 30 08:14:11 crc kubenswrapper[4810]: I0930 08:14:11.371374 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27636f2422460fa31d9d450f77dc80075b12edc005b84fc0a7e19de235732624" Sep 30 08:14:11 crc kubenswrapper[4810]: I0930 08:14:11.371515 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.508642 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-bk866"] Sep 30 08:14:13 crc kubenswrapper[4810]: E0930 08:14:13.508872 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df44374d-4490-461f-aeed-05fae8096705" containerName="util" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.508885 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="df44374d-4490-461f-aeed-05fae8096705" containerName="util" Sep 30 08:14:13 crc kubenswrapper[4810]: E0930 08:14:13.508896 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df44374d-4490-461f-aeed-05fae8096705" containerName="pull" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.508902 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="df44374d-4490-461f-aeed-05fae8096705" containerName="pull" Sep 30 08:14:13 crc kubenswrapper[4810]: E0930 08:14:13.508912 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df44374d-4490-461f-aeed-05fae8096705" containerName="extract" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.508919 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="df44374d-4490-461f-aeed-05fae8096705" containerName="extract" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.509008 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="df44374d-4490-461f-aeed-05fae8096705" containerName="extract" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.509399 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-bk866" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.512107 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.512146 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-rh7mf" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.512459 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.520806 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-bk866"] Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.635886 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7cr7\" (UniqueName: \"kubernetes.io/projected/5a1ec2d4-00e6-4db8-95c0-3079d24f18a8-kube-api-access-n7cr7\") pod \"nmstate-operator-5d6f6cfd66-bk866\" (UID: \"5a1ec2d4-00e6-4db8-95c0-3079d24f18a8\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-bk866" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.737557 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7cr7\" (UniqueName: \"kubernetes.io/projected/5a1ec2d4-00e6-4db8-95c0-3079d24f18a8-kube-api-access-n7cr7\") pod \"nmstate-operator-5d6f6cfd66-bk866\" (UID: \"5a1ec2d4-00e6-4db8-95c0-3079d24f18a8\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-bk866" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.764139 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7cr7\" (UniqueName: \"kubernetes.io/projected/5a1ec2d4-00e6-4db8-95c0-3079d24f18a8-kube-api-access-n7cr7\") pod \"nmstate-operator-5d6f6cfd66-bk866\" (UID: \"5a1ec2d4-00e6-4db8-95c0-3079d24f18a8\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-bk866" Sep 30 08:14:13 crc kubenswrapper[4810]: I0930 08:14:13.822886 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-bk866" Sep 30 08:14:14 crc kubenswrapper[4810]: I0930 08:14:14.156080 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-bk866"] Sep 30 08:14:14 crc kubenswrapper[4810]: I0930 08:14:14.397116 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-bk866" event={"ID":"5a1ec2d4-00e6-4db8-95c0-3079d24f18a8","Type":"ContainerStarted","Data":"9670ebf2e2a0bd8e795942b20ff6c38b989d70c7f72569fc60b540ceb6277708"} Sep 30 08:14:16 crc kubenswrapper[4810]: I0930 08:14:16.418403 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-bk866" event={"ID":"5a1ec2d4-00e6-4db8-95c0-3079d24f18a8","Type":"ContainerStarted","Data":"77e9dbc56f52032d0dbe8c2945ad10219ae4bd2c4b26899d78cb0fba1d20a479"} Sep 30 08:14:17 crc kubenswrapper[4810]: I0930 08:14:17.458985 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-bk866" podStartSLOduration=2.454937518 podStartE2EDuration="4.458954361s" podCreationTimestamp="2025-09-30 08:14:13 +0000 UTC" firstStartedPulling="2025-09-30 08:14:14.169810131 +0000 UTC m=+677.622009408" lastFinishedPulling="2025-09-30 08:14:16.173826974 +0000 UTC m=+679.626026251" observedRunningTime="2025-09-30 08:14:17.451296247 +0000 UTC m=+680.903495554" watchObservedRunningTime="2025-09-30 08:14:17.458954361 +0000 UTC m=+680.911153668" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.467018 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4"] Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.467938 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.470040 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-wfvkp" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.482464 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4"] Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.497665 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm"] Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.500402 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.504013 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.516556 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm"] Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.517330 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hpbb\" (UniqueName: \"kubernetes.io/projected/8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb-kube-api-access-7hpbb\") pod \"nmstate-webhook-6d689559c5-h2zdm\" (UID: \"8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.517376 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6wj4\" (UniqueName: \"kubernetes.io/projected/303e8028-1bcf-4fa7-8314-2ed52bec230a-kube-api-access-f6wj4\") pod \"nmstate-metrics-58fcddf996-qzvw4\" (UID: \"303e8028-1bcf-4fa7-8314-2ed52bec230a\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.517451 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h2zdm\" (UID: \"8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.533323 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-w4dnr"] Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.534992 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.618423 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8lz2\" (UniqueName: \"kubernetes.io/projected/b8df67fa-d671-4174-8306-61127ae22268-kube-api-access-z8lz2\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.618678 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h2zdm\" (UID: \"8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.618790 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b8df67fa-d671-4174-8306-61127ae22268-dbus-socket\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.618889 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b8df67fa-d671-4174-8306-61127ae22268-ovs-socket\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: E0930 08:14:18.618913 4810 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Sep 30 08:14:18 crc kubenswrapper[4810]: E0930 08:14:18.619011 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb-tls-key-pair podName:8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb nodeName:}" failed. No retries permitted until 2025-09-30 08:14:19.118990868 +0000 UTC m=+682.571190135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb-tls-key-pair") pod "nmstate-webhook-6d689559c5-h2zdm" (UID: "8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb") : secret "openshift-nmstate-webhook" not found Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.618923 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hpbb\" (UniqueName: \"kubernetes.io/projected/8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb-kube-api-access-7hpbb\") pod \"nmstate-webhook-6d689559c5-h2zdm\" (UID: \"8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.619137 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6wj4\" (UniqueName: \"kubernetes.io/projected/303e8028-1bcf-4fa7-8314-2ed52bec230a-kube-api-access-f6wj4\") pod \"nmstate-metrics-58fcddf996-qzvw4\" (UID: \"303e8028-1bcf-4fa7-8314-2ed52bec230a\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.619226 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b8df67fa-d671-4174-8306-61127ae22268-nmstate-lock\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.639203 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7"] Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.640133 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.644173 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.644761 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-dhkqq" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.647259 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hpbb\" (UniqueName: \"kubernetes.io/projected/8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb-kube-api-access-7hpbb\") pod \"nmstate-webhook-6d689559c5-h2zdm\" (UID: \"8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.647339 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.656100 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7"] Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.660534 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6wj4\" (UniqueName: \"kubernetes.io/projected/303e8028-1bcf-4fa7-8314-2ed52bec230a-kube-api-access-f6wj4\") pod \"nmstate-metrics-58fcddf996-qzvw4\" (UID: \"303e8028-1bcf-4fa7-8314-2ed52bec230a\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.720106 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b8df67fa-d671-4174-8306-61127ae22268-dbus-socket\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.720141 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b8df67fa-d671-4174-8306-61127ae22268-ovs-socket\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.720173 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b672b0bf-c5cc-467a-9014-99a3269488e8-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-pdhp7\" (UID: \"b672b0bf-c5cc-467a-9014-99a3269488e8\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.720193 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5rhk\" (UniqueName: \"kubernetes.io/projected/b672b0bf-c5cc-467a-9014-99a3269488e8-kube-api-access-k5rhk\") pod \"nmstate-console-plugin-864bb6dfb5-pdhp7\" (UID: \"b672b0bf-c5cc-467a-9014-99a3269488e8\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.720215 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b8df67fa-d671-4174-8306-61127ae22268-nmstate-lock\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.720246 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8lz2\" (UniqueName: \"kubernetes.io/projected/b8df67fa-d671-4174-8306-61127ae22268-kube-api-access-z8lz2\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.720289 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b672b0bf-c5cc-467a-9014-99a3269488e8-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-pdhp7\" (UID: \"b672b0bf-c5cc-467a-9014-99a3269488e8\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.720518 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b8df67fa-d671-4174-8306-61127ae22268-dbus-socket\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.720544 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b8df67fa-d671-4174-8306-61127ae22268-ovs-socket\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.720571 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b8df67fa-d671-4174-8306-61127ae22268-nmstate-lock\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.747334 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8lz2\" (UniqueName: \"kubernetes.io/projected/b8df67fa-d671-4174-8306-61127ae22268-kube-api-access-z8lz2\") pod \"nmstate-handler-w4dnr\" (UID: \"b8df67fa-d671-4174-8306-61127ae22268\") " pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.792748 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.820961 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6494bbbb7b-h42pk"] Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.821064 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b672b0bf-c5cc-467a-9014-99a3269488e8-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-pdhp7\" (UID: \"b672b0bf-c5cc-467a-9014-99a3269488e8\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.821483 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5rhk\" (UniqueName: \"kubernetes.io/projected/b672b0bf-c5cc-467a-9014-99a3269488e8-kube-api-access-k5rhk\") pod \"nmstate-console-plugin-864bb6dfb5-pdhp7\" (UID: \"b672b0bf-c5cc-467a-9014-99a3269488e8\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.821526 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b672b0bf-c5cc-467a-9014-99a3269488e8-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-pdhp7\" (UID: \"b672b0bf-c5cc-467a-9014-99a3269488e8\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.821827 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b672b0bf-c5cc-467a-9014-99a3269488e8-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-pdhp7\" (UID: \"b672b0bf-c5cc-467a-9014-99a3269488e8\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.822054 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.829908 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b672b0bf-c5cc-467a-9014-99a3269488e8-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-pdhp7\" (UID: \"b672b0bf-c5cc-467a-9014-99a3269488e8\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.837583 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6494bbbb7b-h42pk"] Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.851134 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5rhk\" (UniqueName: \"kubernetes.io/projected/b672b0bf-c5cc-467a-9014-99a3269488e8-kube-api-access-k5rhk\") pod \"nmstate-console-plugin-864bb6dfb5-pdhp7\" (UID: \"b672b0bf-c5cc-467a-9014-99a3269488e8\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.857400 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:18 crc kubenswrapper[4810]: W0930 08:14:18.878816 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8df67fa_d671_4174_8306_61127ae22268.slice/crio-7628275d01da4928ebdae40b0860f55312074ffb1873d34638297a8029911a74 WatchSource:0}: Error finding container 7628275d01da4928ebdae40b0860f55312074ffb1873d34638297a8029911a74: Status 404 returned error can't find the container with id 7628275d01da4928ebdae40b0860f55312074ffb1873d34638297a8029911a74 Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.922755 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcssf\" (UniqueName: \"kubernetes.io/projected/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-kube-api-access-lcssf\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.922798 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-console-serving-cert\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.922830 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-service-ca\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.922869 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-console-oauth-config\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.922888 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-oauth-serving-cert\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.922923 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-trusted-ca-bundle\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.922940 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-console-config\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:18 crc kubenswrapper[4810]: I0930 08:14:18.991200 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.026967 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-service-ca\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.027041 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-console-oauth-config\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.027063 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-oauth-serving-cert\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.027867 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-trusted-ca-bundle\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.027952 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-console-config\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.027998 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcssf\" (UniqueName: \"kubernetes.io/projected/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-kube-api-access-lcssf\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.028059 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-console-serving-cert\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.029472 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-oauth-serving-cert\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.030040 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-service-ca\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.032681 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-console-oauth-config\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.033311 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-console-config\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.033409 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-trusted-ca-bundle\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.035913 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-console-serving-cert\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.045085 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4"] Sep 30 08:14:19 crc kubenswrapper[4810]: W0930 08:14:19.051851 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod303e8028_1bcf_4fa7_8314_2ed52bec230a.slice/crio-5f733bc8d1e49d02db7a2dac9e377f08d9b1401e559d609b006d40affc8c248b WatchSource:0}: Error finding container 5f733bc8d1e49d02db7a2dac9e377f08d9b1401e559d609b006d40affc8c248b: Status 404 returned error can't find the container with id 5f733bc8d1e49d02db7a2dac9e377f08d9b1401e559d609b006d40affc8c248b Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.052869 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcssf\" (UniqueName: \"kubernetes.io/projected/dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c-kube-api-access-lcssf\") pod \"console-6494bbbb7b-h42pk\" (UID: \"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c\") " pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.129009 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h2zdm\" (UID: \"8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.133204 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h2zdm\" (UID: \"8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.171768 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.209414 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7"] Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.418833 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6494bbbb7b-h42pk"] Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.421149 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" Sep 30 08:14:19 crc kubenswrapper[4810]: W0930 08:14:19.424872 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbb7d8d9_015a_4ff5_8b09_b0a4e4616f4c.slice/crio-9897203e8897ecfbf09cfc162ffd73c9fed9e9e54ee5154e8a332f7b22b313d8 WatchSource:0}: Error finding container 9897203e8897ecfbf09cfc162ffd73c9fed9e9e54ee5154e8a332f7b22b313d8: Status 404 returned error can't find the container with id 9897203e8897ecfbf09cfc162ffd73c9fed9e9e54ee5154e8a332f7b22b313d8 Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.441527 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6494bbbb7b-h42pk" event={"ID":"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c","Type":"ContainerStarted","Data":"9897203e8897ecfbf09cfc162ffd73c9fed9e9e54ee5154e8a332f7b22b313d8"} Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.443993 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" event={"ID":"b672b0bf-c5cc-467a-9014-99a3269488e8","Type":"ContainerStarted","Data":"b94ef58c5d6737864dd90ccb52cbc1610d63de54bea68f27bad9a1f69efb6e05"} Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.451731 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-w4dnr" event={"ID":"b8df67fa-d671-4174-8306-61127ae22268","Type":"ContainerStarted","Data":"7628275d01da4928ebdae40b0860f55312074ffb1873d34638297a8029911a74"} Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.454461 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4" event={"ID":"303e8028-1bcf-4fa7-8314-2ed52bec230a","Type":"ContainerStarted","Data":"5f733bc8d1e49d02db7a2dac9e377f08d9b1401e559d609b006d40affc8c248b"} Sep 30 08:14:19 crc kubenswrapper[4810]: I0930 08:14:19.697444 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm"] Sep 30 08:14:19 crc kubenswrapper[4810]: W0930 08:14:19.706333 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dbcd62d_c15f_4d23_ac4a_80c8eda6dfcb.slice/crio-dbf9f57b5bb2ebb01cbdbe731cb0fe0a57ab6b999d793e0d4fd85c649b89c84f WatchSource:0}: Error finding container dbf9f57b5bb2ebb01cbdbe731cb0fe0a57ab6b999d793e0d4fd85c649b89c84f: Status 404 returned error can't find the container with id dbf9f57b5bb2ebb01cbdbe731cb0fe0a57ab6b999d793e0d4fd85c649b89c84f Sep 30 08:14:20 crc kubenswrapper[4810]: I0930 08:14:20.463185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6494bbbb7b-h42pk" event={"ID":"dbb7d8d9-015a-4ff5-8b09-b0a4e4616f4c","Type":"ContainerStarted","Data":"f01e58e4ba1fe369e4360306245c2fa38d988ace214643c80e4fabbb37bc1751"} Sep 30 08:14:20 crc kubenswrapper[4810]: I0930 08:14:20.465978 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" event={"ID":"8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb","Type":"ContainerStarted","Data":"dbf9f57b5bb2ebb01cbdbe731cb0fe0a57ab6b999d793e0d4fd85c649b89c84f"} Sep 30 08:14:21 crc kubenswrapper[4810]: I0930 08:14:21.478801 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4" event={"ID":"303e8028-1bcf-4fa7-8314-2ed52bec230a","Type":"ContainerStarted","Data":"1d772953578d7e14f644093ca241b6456281bae160226ecfa145a0d3ee7fd50e"} Sep 30 08:14:21 crc kubenswrapper[4810]: I0930 08:14:21.481171 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" event={"ID":"8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb","Type":"ContainerStarted","Data":"a85aefb59177e1b849902ed6eb25274ed58978aa20d73e8bbd166587bcc50217"} Sep 30 08:14:21 crc kubenswrapper[4810]: I0930 08:14:21.481694 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" Sep 30 08:14:21 crc kubenswrapper[4810]: I0930 08:14:21.501542 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6494bbbb7b-h42pk" podStartSLOduration=3.50152225 podStartE2EDuration="3.50152225s" podCreationTimestamp="2025-09-30 08:14:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:14:20.484366464 +0000 UTC m=+683.936565731" watchObservedRunningTime="2025-09-30 08:14:21.50152225 +0000 UTC m=+684.953721517" Sep 30 08:14:22 crc kubenswrapper[4810]: I0930 08:14:22.496165 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-w4dnr" event={"ID":"b8df67fa-d671-4174-8306-61127ae22268","Type":"ContainerStarted","Data":"2a314976f15c93c1ca3594378a4afcddb7436c23da5e4a2ef51a82363413d3f1"} Sep 30 08:14:22 crc kubenswrapper[4810]: I0930 08:14:22.521111 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" podStartSLOduration=2.978655688 podStartE2EDuration="4.521075494s" podCreationTimestamp="2025-09-30 08:14:18 +0000 UTC" firstStartedPulling="2025-09-30 08:14:19.70919628 +0000 UTC m=+683.161395547" lastFinishedPulling="2025-09-30 08:14:21.251616076 +0000 UTC m=+684.703815353" observedRunningTime="2025-09-30 08:14:21.500616993 +0000 UTC m=+684.952816270" watchObservedRunningTime="2025-09-30 08:14:22.521075494 +0000 UTC m=+685.973274801" Sep 30 08:14:23 crc kubenswrapper[4810]: I0930 08:14:23.506561 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" event={"ID":"b672b0bf-c5cc-467a-9014-99a3269488e8","Type":"ContainerStarted","Data":"aa48a4256372ec5bddfa717dd5d7a580ca12774340a07c20bc41256644535b9c"} Sep 30 08:14:23 crc kubenswrapper[4810]: I0930 08:14:23.507014 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:23 crc kubenswrapper[4810]: I0930 08:14:23.522898 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pdhp7" podStartSLOduration=2.370745668 podStartE2EDuration="5.52287611s" podCreationTimestamp="2025-09-30 08:14:18 +0000 UTC" firstStartedPulling="2025-09-30 08:14:19.226596777 +0000 UTC m=+682.678796044" lastFinishedPulling="2025-09-30 08:14:22.378727189 +0000 UTC m=+685.830926486" observedRunningTime="2025-09-30 08:14:23.522769377 +0000 UTC m=+686.974968644" watchObservedRunningTime="2025-09-30 08:14:23.52287611 +0000 UTC m=+686.975075377" Sep 30 08:14:23 crc kubenswrapper[4810]: I0930 08:14:23.526803 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-w4dnr" podStartSLOduration=3.123013683 podStartE2EDuration="5.526783535s" podCreationTimestamp="2025-09-30 08:14:18 +0000 UTC" firstStartedPulling="2025-09-30 08:14:18.888171774 +0000 UTC m=+682.340371041" lastFinishedPulling="2025-09-30 08:14:21.291941616 +0000 UTC m=+684.744140893" observedRunningTime="2025-09-30 08:14:22.520212019 +0000 UTC m=+685.972411296" watchObservedRunningTime="2025-09-30 08:14:23.526783535 +0000 UTC m=+686.978982812" Sep 30 08:14:24 crc kubenswrapper[4810]: I0930 08:14:24.517227 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4" event={"ID":"303e8028-1bcf-4fa7-8314-2ed52bec230a","Type":"ContainerStarted","Data":"33651cc5b2f3ed3fca8149bd57e86f0c9ffa3cd699cd527bb3c4ece39dd0bff7"} Sep 30 08:14:24 crc kubenswrapper[4810]: I0930 08:14:24.542755 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58fcddf996-qzvw4" podStartSLOduration=1.854586295 podStartE2EDuration="6.542726895s" podCreationTimestamp="2025-09-30 08:14:18 +0000 UTC" firstStartedPulling="2025-09-30 08:14:19.054329326 +0000 UTC m=+682.506528593" lastFinishedPulling="2025-09-30 08:14:23.742469916 +0000 UTC m=+687.194669193" observedRunningTime="2025-09-30 08:14:24.537903623 +0000 UTC m=+687.990102960" watchObservedRunningTime="2025-09-30 08:14:24.542726895 +0000 UTC m=+687.994926202" Sep 30 08:14:28 crc kubenswrapper[4810]: I0930 08:14:28.900662 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-w4dnr" Sep 30 08:14:29 crc kubenswrapper[4810]: I0930 08:14:29.172597 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:29 crc kubenswrapper[4810]: I0930 08:14:29.173177 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:29 crc kubenswrapper[4810]: I0930 08:14:29.181028 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:29 crc kubenswrapper[4810]: I0930 08:14:29.559366 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6494bbbb7b-h42pk" Sep 30 08:14:29 crc kubenswrapper[4810]: I0930 08:14:29.675200 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5w577"] Sep 30 08:14:39 crc kubenswrapper[4810]: I0930 08:14:39.432162 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h2zdm" Sep 30 08:14:54 crc kubenswrapper[4810]: I0930 08:14:54.723070 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-5w577" podUID="64fd38ac-d0ef-430c-aed5-5a4e4944986c" containerName="console" containerID="cri-o://9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda" gracePeriod=15 Sep 30 08:14:54 crc kubenswrapper[4810]: I0930 08:14:54.859329 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8"] Sep 30 08:14:54 crc kubenswrapper[4810]: I0930 08:14:54.862229 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:14:54 crc kubenswrapper[4810]: I0930 08:14:54.865765 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 30 08:14:54 crc kubenswrapper[4810]: I0930 08:14:54.872023 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8"] Sep 30 08:14:54 crc kubenswrapper[4810]: I0930 08:14:54.958844 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7hrd\" (UniqueName: \"kubernetes.io/projected/cee1f0d2-0013-4803-b758-f3efb50a80b5-kube-api-access-m7hrd\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:14:54 crc kubenswrapper[4810]: I0930 08:14:54.958898 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:14:54 crc kubenswrapper[4810]: I0930 08:14:54.958959 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.060106 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7hrd\" (UniqueName: \"kubernetes.io/projected/cee1f0d2-0013-4803-b758-f3efb50a80b5-kube-api-access-m7hrd\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.060179 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.060256 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.061056 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.061096 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.081405 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7hrd\" (UniqueName: \"kubernetes.io/projected/cee1f0d2-0013-4803-b758-f3efb50a80b5-kube-api-access-m7hrd\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.175939 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5w577_64fd38ac-d0ef-430c-aed5-5a4e4944986c/console/0.log" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.176026 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.228374 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.363510 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-oauth-config\") pod \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.363788 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cbfn\" (UniqueName: \"kubernetes.io/projected/64fd38ac-d0ef-430c-aed5-5a4e4944986c-kube-api-access-2cbfn\") pod \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.363824 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-service-ca\") pod \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.363862 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-serving-cert\") pod \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.363887 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-config\") pod \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.363917 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-trusted-ca-bundle\") pod \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.363940 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-oauth-serving-cert\") pod \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\" (UID: \"64fd38ac-d0ef-430c-aed5-5a4e4944986c\") " Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.364929 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "64fd38ac-d0ef-430c-aed5-5a4e4944986c" (UID: "64fd38ac-d0ef-430c-aed5-5a4e4944986c"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.365526 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-service-ca" (OuterVolumeSpecName: "service-ca") pod "64fd38ac-d0ef-430c-aed5-5a4e4944986c" (UID: "64fd38ac-d0ef-430c-aed5-5a4e4944986c"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.365594 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-config" (OuterVolumeSpecName: "console-config") pod "64fd38ac-d0ef-430c-aed5-5a4e4944986c" (UID: "64fd38ac-d0ef-430c-aed5-5a4e4944986c"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.365827 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "64fd38ac-d0ef-430c-aed5-5a4e4944986c" (UID: "64fd38ac-d0ef-430c-aed5-5a4e4944986c"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.367941 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "64fd38ac-d0ef-430c-aed5-5a4e4944986c" (UID: "64fd38ac-d0ef-430c-aed5-5a4e4944986c"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.368687 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64fd38ac-d0ef-430c-aed5-5a4e4944986c-kube-api-access-2cbfn" (OuterVolumeSpecName: "kube-api-access-2cbfn") pod "64fd38ac-d0ef-430c-aed5-5a4e4944986c" (UID: "64fd38ac-d0ef-430c-aed5-5a4e4944986c"). InnerVolumeSpecName "kube-api-access-2cbfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.370174 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "64fd38ac-d0ef-430c-aed5-5a4e4944986c" (UID: "64fd38ac-d0ef-430c-aed5-5a4e4944986c"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.453991 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8"] Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.466229 4810 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.466286 4810 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.466300 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cbfn\" (UniqueName: \"kubernetes.io/projected/64fd38ac-d0ef-430c-aed5-5a4e4944986c-kube-api-access-2cbfn\") on node \"crc\" DevicePath \"\"" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.466312 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.466325 4810 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.466335 4810 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-console-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.466345 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64fd38ac-d0ef-430c-aed5-5a4e4944986c-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.756606 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5w577_64fd38ac-d0ef-430c-aed5-5a4e4944986c/console/0.log" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.756915 4810 generic.go:334] "Generic (PLEG): container finished" podID="64fd38ac-d0ef-430c-aed5-5a4e4944986c" containerID="9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda" exitCode=2 Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.756982 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5w577" event={"ID":"64fd38ac-d0ef-430c-aed5-5a4e4944986c","Type":"ContainerDied","Data":"9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda"} Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.757014 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5w577" event={"ID":"64fd38ac-d0ef-430c-aed5-5a4e4944986c","Type":"ContainerDied","Data":"030ee37eb2e6e0b30ca1099988ce7322f75b321380bcb63716be10c79f73e47d"} Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.757033 4810 scope.go:117] "RemoveContainer" containerID="9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.757034 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5w577" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.760669 4810 generic.go:334] "Generic (PLEG): container finished" podID="cee1f0d2-0013-4803-b758-f3efb50a80b5" containerID="b1f5ce10fe78cb82a29b009e6692c519970ba46b1a50f67d6234af877f3aa64a" exitCode=0 Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.760735 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" event={"ID":"cee1f0d2-0013-4803-b758-f3efb50a80b5","Type":"ContainerDied","Data":"b1f5ce10fe78cb82a29b009e6692c519970ba46b1a50f67d6234af877f3aa64a"} Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.760779 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" event={"ID":"cee1f0d2-0013-4803-b758-f3efb50a80b5","Type":"ContainerStarted","Data":"1dfd04ced339c069daa9a9e0a748a5256ab0768ffd15499920cf2a43307b5b69"} Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.780820 4810 scope.go:117] "RemoveContainer" containerID="9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda" Sep 30 08:14:55 crc kubenswrapper[4810]: E0930 08:14:55.781438 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda\": container with ID starting with 9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda not found: ID does not exist" containerID="9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.781479 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda"} err="failed to get container status \"9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda\": rpc error: code = NotFound desc = could not find container \"9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda\": container with ID starting with 9b1e3e786f7568ed15664a7f47106ae7ae967036e0819feb9cbd8f0957677cda not found: ID does not exist" Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.804030 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5w577"] Sep 30 08:14:55 crc kubenswrapper[4810]: I0930 08:14:55.807667 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-5w577"] Sep 30 08:14:57 crc kubenswrapper[4810]: I0930 08:14:57.317726 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64fd38ac-d0ef-430c-aed5-5a4e4944986c" path="/var/lib/kubelet/pods/64fd38ac-d0ef-430c-aed5-5a4e4944986c/volumes" Sep 30 08:14:58 crc kubenswrapper[4810]: I0930 08:14:58.786386 4810 generic.go:334] "Generic (PLEG): container finished" podID="cee1f0d2-0013-4803-b758-f3efb50a80b5" containerID="1bcf0ecae07667b9600afcbef1b7f15bb98804ebf8ff74d26e36069c2193050c" exitCode=0 Sep 30 08:14:58 crc kubenswrapper[4810]: I0930 08:14:58.786517 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" event={"ID":"cee1f0d2-0013-4803-b758-f3efb50a80b5","Type":"ContainerDied","Data":"1bcf0ecae07667b9600afcbef1b7f15bb98804ebf8ff74d26e36069c2193050c"} Sep 30 08:14:59 crc kubenswrapper[4810]: I0930 08:14:59.799013 4810 generic.go:334] "Generic (PLEG): container finished" podID="cee1f0d2-0013-4803-b758-f3efb50a80b5" containerID="710580a1b68e933b64dae86043153e2e6e71efb0ae72726ff5f5b7c0a28fc03e" exitCode=0 Sep 30 08:14:59 crc kubenswrapper[4810]: I0930 08:14:59.799079 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" event={"ID":"cee1f0d2-0013-4803-b758-f3efb50a80b5","Type":"ContainerDied","Data":"710580a1b68e933b64dae86043153e2e6e71efb0ae72726ff5f5b7c0a28fc03e"} Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.139497 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8"] Sep 30 08:15:00 crc kubenswrapper[4810]: E0930 08:15:00.139779 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64fd38ac-d0ef-430c-aed5-5a4e4944986c" containerName="console" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.139793 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="64fd38ac-d0ef-430c-aed5-5a4e4944986c" containerName="console" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.139907 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="64fd38ac-d0ef-430c-aed5-5a4e4944986c" containerName="console" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.140441 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.143252 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.143965 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.156055 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8"] Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.332534 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvm6c\" (UniqueName: \"kubernetes.io/projected/9ce63664-583b-408a-a9d1-061552873891-kube-api-access-vvm6c\") pod \"collect-profiles-29320335-9wnd8\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.332595 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ce63664-583b-408a-a9d1-061552873891-config-volume\") pod \"collect-profiles-29320335-9wnd8\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.332680 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ce63664-583b-408a-a9d1-061552873891-secret-volume\") pod \"collect-profiles-29320335-9wnd8\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.434584 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvm6c\" (UniqueName: \"kubernetes.io/projected/9ce63664-583b-408a-a9d1-061552873891-kube-api-access-vvm6c\") pod \"collect-profiles-29320335-9wnd8\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.434666 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ce63664-583b-408a-a9d1-061552873891-config-volume\") pod \"collect-profiles-29320335-9wnd8\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.434790 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ce63664-583b-408a-a9d1-061552873891-secret-volume\") pod \"collect-profiles-29320335-9wnd8\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.436949 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ce63664-583b-408a-a9d1-061552873891-config-volume\") pod \"collect-profiles-29320335-9wnd8\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.443533 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ce63664-583b-408a-a9d1-061552873891-secret-volume\") pod \"collect-profiles-29320335-9wnd8\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.459017 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvm6c\" (UniqueName: \"kubernetes.io/projected/9ce63664-583b-408a-a9d1-061552873891-kube-api-access-vvm6c\") pod \"collect-profiles-29320335-9wnd8\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:00 crc kubenswrapper[4810]: I0930 08:15:00.758895 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.048786 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8"] Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.089253 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.145919 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7hrd\" (UniqueName: \"kubernetes.io/projected/cee1f0d2-0013-4803-b758-f3efb50a80b5-kube-api-access-m7hrd\") pod \"cee1f0d2-0013-4803-b758-f3efb50a80b5\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.145982 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-util\") pod \"cee1f0d2-0013-4803-b758-f3efb50a80b5\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.146018 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-bundle\") pod \"cee1f0d2-0013-4803-b758-f3efb50a80b5\" (UID: \"cee1f0d2-0013-4803-b758-f3efb50a80b5\") " Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.147483 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-bundle" (OuterVolumeSpecName: "bundle") pod "cee1f0d2-0013-4803-b758-f3efb50a80b5" (UID: "cee1f0d2-0013-4803-b758-f3efb50a80b5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.153736 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee1f0d2-0013-4803-b758-f3efb50a80b5-kube-api-access-m7hrd" (OuterVolumeSpecName: "kube-api-access-m7hrd") pod "cee1f0d2-0013-4803-b758-f3efb50a80b5" (UID: "cee1f0d2-0013-4803-b758-f3efb50a80b5"). InnerVolumeSpecName "kube-api-access-m7hrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.247357 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.247419 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7hrd\" (UniqueName: \"kubernetes.io/projected/cee1f0d2-0013-4803-b758-f3efb50a80b5-kube-api-access-m7hrd\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.472944 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-util" (OuterVolumeSpecName: "util") pod "cee1f0d2-0013-4803-b758-f3efb50a80b5" (UID: "cee1f0d2-0013-4803-b758-f3efb50a80b5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.552723 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cee1f0d2-0013-4803-b758-f3efb50a80b5-util\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.818803 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" event={"ID":"9ce63664-583b-408a-a9d1-061552873891","Type":"ContainerStarted","Data":"4a8ae09deb8c59bb0dec3236ae6e7affa65036022541214f402cfa5168f3ddf1"} Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.818873 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" event={"ID":"9ce63664-583b-408a-a9d1-061552873891","Type":"ContainerStarted","Data":"455ff31cd1e417c3c854be08c139a1d6534562cd09ee04f3a4f150a59a51dc7f"} Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.823716 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" event={"ID":"cee1f0d2-0013-4803-b758-f3efb50a80b5","Type":"ContainerDied","Data":"1dfd04ced339c069daa9a9e0a748a5256ab0768ffd15499920cf2a43307b5b69"} Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.823772 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dfd04ced339c069daa9a9e0a748a5256ab0768ffd15499920cf2a43307b5b69" Sep 30 08:15:01 crc kubenswrapper[4810]: I0930 08:15:01.823801 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8" Sep 30 08:15:02 crc kubenswrapper[4810]: I0930 08:15:02.833474 4810 generic.go:334] "Generic (PLEG): container finished" podID="9ce63664-583b-408a-a9d1-061552873891" containerID="4a8ae09deb8c59bb0dec3236ae6e7affa65036022541214f402cfa5168f3ddf1" exitCode=0 Sep 30 08:15:02 crc kubenswrapper[4810]: I0930 08:15:02.833592 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" event={"ID":"9ce63664-583b-408a-a9d1-061552873891","Type":"ContainerDied","Data":"4a8ae09deb8c59bb0dec3236ae6e7affa65036022541214f402cfa5168f3ddf1"} Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.103486 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.286643 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ce63664-583b-408a-a9d1-061552873891-secret-volume\") pod \"9ce63664-583b-408a-a9d1-061552873891\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.286740 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ce63664-583b-408a-a9d1-061552873891-config-volume\") pod \"9ce63664-583b-408a-a9d1-061552873891\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.286789 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvm6c\" (UniqueName: \"kubernetes.io/projected/9ce63664-583b-408a-a9d1-061552873891-kube-api-access-vvm6c\") pod \"9ce63664-583b-408a-a9d1-061552873891\" (UID: \"9ce63664-583b-408a-a9d1-061552873891\") " Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.287517 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce63664-583b-408a-a9d1-061552873891-config-volume" (OuterVolumeSpecName: "config-volume") pod "9ce63664-583b-408a-a9d1-061552873891" (UID: "9ce63664-583b-408a-a9d1-061552873891"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.294675 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ce63664-583b-408a-a9d1-061552873891-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9ce63664-583b-408a-a9d1-061552873891" (UID: "9ce63664-583b-408a-a9d1-061552873891"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.308263 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ce63664-583b-408a-a9d1-061552873891-kube-api-access-vvm6c" (OuterVolumeSpecName: "kube-api-access-vvm6c") pod "9ce63664-583b-408a-a9d1-061552873891" (UID: "9ce63664-583b-408a-a9d1-061552873891"). InnerVolumeSpecName "kube-api-access-vvm6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.388052 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ce63664-583b-408a-a9d1-061552873891-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.388127 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ce63664-583b-408a-a9d1-061552873891-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.388229 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvm6c\" (UniqueName: \"kubernetes.io/projected/9ce63664-583b-408a-a9d1-061552873891-kube-api-access-vvm6c\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.847225 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" event={"ID":"9ce63664-583b-408a-a9d1-061552873891","Type":"ContainerDied","Data":"455ff31cd1e417c3c854be08c139a1d6534562cd09ee04f3a4f150a59a51dc7f"} Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.847301 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="455ff31cd1e417c3c854be08c139a1d6534562cd09ee04f3a4f150a59a51dc7f" Sep 30 08:15:04 crc kubenswrapper[4810]: I0930 08:15:04.847351 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.140537 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5"] Sep 30 08:15:10 crc kubenswrapper[4810]: E0930 08:15:10.141414 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee1f0d2-0013-4803-b758-f3efb50a80b5" containerName="pull" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.141429 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee1f0d2-0013-4803-b758-f3efb50a80b5" containerName="pull" Sep 30 08:15:10 crc kubenswrapper[4810]: E0930 08:15:10.141441 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce63664-583b-408a-a9d1-061552873891" containerName="collect-profiles" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.141447 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce63664-583b-408a-a9d1-061552873891" containerName="collect-profiles" Sep 30 08:15:10 crc kubenswrapper[4810]: E0930 08:15:10.141461 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee1f0d2-0013-4803-b758-f3efb50a80b5" containerName="util" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.141467 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee1f0d2-0013-4803-b758-f3efb50a80b5" containerName="util" Sep 30 08:15:10 crc kubenswrapper[4810]: E0930 08:15:10.141478 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee1f0d2-0013-4803-b758-f3efb50a80b5" containerName="extract" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.141483 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee1f0d2-0013-4803-b758-f3efb50a80b5" containerName="extract" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.141580 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ce63664-583b-408a-a9d1-061552873891" containerName="collect-profiles" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.141593 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee1f0d2-0013-4803-b758-f3efb50a80b5" containerName="extract" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.141986 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.146126 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.146143 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.146292 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.146462 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.146478 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-r6lk6" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.155791 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5"] Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.271867 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d9ct\" (UniqueName: \"kubernetes.io/projected/9710139d-e7f3-4921-b84a-4b6525891f99-kube-api-access-7d9ct\") pod \"metallb-operator-controller-manager-74b66975c5-rtrd5\" (UID: \"9710139d-e7f3-4921-b84a-4b6525891f99\") " pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.271931 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9710139d-e7f3-4921-b84a-4b6525891f99-webhook-cert\") pod \"metallb-operator-controller-manager-74b66975c5-rtrd5\" (UID: \"9710139d-e7f3-4921-b84a-4b6525891f99\") " pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.272138 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9710139d-e7f3-4921-b84a-4b6525891f99-apiservice-cert\") pod \"metallb-operator-controller-manager-74b66975c5-rtrd5\" (UID: \"9710139d-e7f3-4921-b84a-4b6525891f99\") " pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.373154 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9710139d-e7f3-4921-b84a-4b6525891f99-apiservice-cert\") pod \"metallb-operator-controller-manager-74b66975c5-rtrd5\" (UID: \"9710139d-e7f3-4921-b84a-4b6525891f99\") " pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.373291 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d9ct\" (UniqueName: \"kubernetes.io/projected/9710139d-e7f3-4921-b84a-4b6525891f99-kube-api-access-7d9ct\") pod \"metallb-operator-controller-manager-74b66975c5-rtrd5\" (UID: \"9710139d-e7f3-4921-b84a-4b6525891f99\") " pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.373341 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9710139d-e7f3-4921-b84a-4b6525891f99-webhook-cert\") pod \"metallb-operator-controller-manager-74b66975c5-rtrd5\" (UID: \"9710139d-e7f3-4921-b84a-4b6525891f99\") " pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.378819 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9710139d-e7f3-4921-b84a-4b6525891f99-apiservice-cert\") pod \"metallb-operator-controller-manager-74b66975c5-rtrd5\" (UID: \"9710139d-e7f3-4921-b84a-4b6525891f99\") " pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.382785 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9710139d-e7f3-4921-b84a-4b6525891f99-webhook-cert\") pod \"metallb-operator-controller-manager-74b66975c5-rtrd5\" (UID: \"9710139d-e7f3-4921-b84a-4b6525891f99\") " pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.392893 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d9ct\" (UniqueName: \"kubernetes.io/projected/9710139d-e7f3-4921-b84a-4b6525891f99-kube-api-access-7d9ct\") pod \"metallb-operator-controller-manager-74b66975c5-rtrd5\" (UID: \"9710139d-e7f3-4921-b84a-4b6525891f99\") " pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.446809 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27"] Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.447574 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.459980 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.460028 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-vmkdb" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.459981 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.460123 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.467904 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27"] Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.473985 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9d49935b-5628-43f0-b527-18d3b55aaa3a-webhook-cert\") pod \"metallb-operator-webhook-server-768c8b4c95-d2v27\" (UID: \"9d49935b-5628-43f0-b527-18d3b55aaa3a\") " pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.474070 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxlxw\" (UniqueName: \"kubernetes.io/projected/9d49935b-5628-43f0-b527-18d3b55aaa3a-kube-api-access-gxlxw\") pod \"metallb-operator-webhook-server-768c8b4c95-d2v27\" (UID: \"9d49935b-5628-43f0-b527-18d3b55aaa3a\") " pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.474125 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9d49935b-5628-43f0-b527-18d3b55aaa3a-apiservice-cert\") pod \"metallb-operator-webhook-server-768c8b4c95-d2v27\" (UID: \"9d49935b-5628-43f0-b527-18d3b55aaa3a\") " pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.575363 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9d49935b-5628-43f0-b527-18d3b55aaa3a-webhook-cert\") pod \"metallb-operator-webhook-server-768c8b4c95-d2v27\" (UID: \"9d49935b-5628-43f0-b527-18d3b55aaa3a\") " pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.575439 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxlxw\" (UniqueName: \"kubernetes.io/projected/9d49935b-5628-43f0-b527-18d3b55aaa3a-kube-api-access-gxlxw\") pod \"metallb-operator-webhook-server-768c8b4c95-d2v27\" (UID: \"9d49935b-5628-43f0-b527-18d3b55aaa3a\") " pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.575468 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9d49935b-5628-43f0-b527-18d3b55aaa3a-apiservice-cert\") pod \"metallb-operator-webhook-server-768c8b4c95-d2v27\" (UID: \"9d49935b-5628-43f0-b527-18d3b55aaa3a\") " pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.579949 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9d49935b-5628-43f0-b527-18d3b55aaa3a-webhook-cert\") pod \"metallb-operator-webhook-server-768c8b4c95-d2v27\" (UID: \"9d49935b-5628-43f0-b527-18d3b55aaa3a\") " pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.595668 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9d49935b-5628-43f0-b527-18d3b55aaa3a-apiservice-cert\") pod \"metallb-operator-webhook-server-768c8b4c95-d2v27\" (UID: \"9d49935b-5628-43f0-b527-18d3b55aaa3a\") " pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.623035 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxlxw\" (UniqueName: \"kubernetes.io/projected/9d49935b-5628-43f0-b527-18d3b55aaa3a-kube-api-access-gxlxw\") pod \"metallb-operator-webhook-server-768c8b4c95-d2v27\" (UID: \"9d49935b-5628-43f0-b527-18d3b55aaa3a\") " pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.722437 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5"] Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.774627 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:10 crc kubenswrapper[4810]: I0930 08:15:10.901662 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" event={"ID":"9710139d-e7f3-4921-b84a-4b6525891f99","Type":"ContainerStarted","Data":"d8ae309282ddf5d85bd614ef124c9d63cd5f8eb28ef6f9f4e150f3a2da906ade"} Sep 30 08:15:11 crc kubenswrapper[4810]: I0930 08:15:11.318662 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27"] Sep 30 08:15:11 crc kubenswrapper[4810]: W0930 08:15:11.323797 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d49935b_5628_43f0_b527_18d3b55aaa3a.slice/crio-7ab5c407f47ace45d22c5c8612186a246feb22864d2c1d217e4004e50d868fe9 WatchSource:0}: Error finding container 7ab5c407f47ace45d22c5c8612186a246feb22864d2c1d217e4004e50d868fe9: Status 404 returned error can't find the container with id 7ab5c407f47ace45d22c5c8612186a246feb22864d2c1d217e4004e50d868fe9 Sep 30 08:15:11 crc kubenswrapper[4810]: I0930 08:15:11.909144 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" event={"ID":"9d49935b-5628-43f0-b527-18d3b55aaa3a","Type":"ContainerStarted","Data":"7ab5c407f47ace45d22c5c8612186a246feb22864d2c1d217e4004e50d868fe9"} Sep 30 08:15:26 crc kubenswrapper[4810]: I0930 08:15:26.590458 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qdcml"] Sep 30 08:15:26 crc kubenswrapper[4810]: I0930 08:15:26.591412 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" podUID="80a2e9d7-a1ec-436d-8634-89c763644dec" containerName="controller-manager" containerID="cri-o://638c48838c29dccb16889232e47ee51085f0be306b2c6857be7f7195f2fa6135" gracePeriod=30 Sep 30 08:15:26 crc kubenswrapper[4810]: I0930 08:15:26.614881 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf"] Sep 30 08:15:26 crc kubenswrapper[4810]: I0930 08:15:26.615163 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" podUID="479ebf74-62ea-4516-8c29-8d9cdf128fb4" containerName="route-controller-manager" containerID="cri-o://2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e" gracePeriod=30 Sep 30 08:15:27 crc kubenswrapper[4810]: E0930 08:15:27.393520 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80a2e9d7_a1ec_436d_8634_89c763644dec.slice/crio-conmon-638c48838c29dccb16889232e47ee51085f0be306b2c6857be7f7195f2fa6135.scope\": RecentStats: unable to find data in memory cache]" Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.654715 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.738581 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/479ebf74-62ea-4516-8c29-8d9cdf128fb4-serving-cert\") pod \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.738666 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wskjk\" (UniqueName: \"kubernetes.io/projected/479ebf74-62ea-4516-8c29-8d9cdf128fb4-kube-api-access-wskjk\") pod \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.738904 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-config\") pod \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.738937 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-client-ca\") pod \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\" (UID: \"479ebf74-62ea-4516-8c29-8d9cdf128fb4\") " Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.739804 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-config" (OuterVolumeSpecName: "config") pod "479ebf74-62ea-4516-8c29-8d9cdf128fb4" (UID: "479ebf74-62ea-4516-8c29-8d9cdf128fb4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.739826 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-client-ca" (OuterVolumeSpecName: "client-ca") pod "479ebf74-62ea-4516-8c29-8d9cdf128fb4" (UID: "479ebf74-62ea-4516-8c29-8d9cdf128fb4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.746545 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/479ebf74-62ea-4516-8c29-8d9cdf128fb4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "479ebf74-62ea-4516-8c29-8d9cdf128fb4" (UID: "479ebf74-62ea-4516-8c29-8d9cdf128fb4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.748306 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/479ebf74-62ea-4516-8c29-8d9cdf128fb4-kube-api-access-wskjk" (OuterVolumeSpecName: "kube-api-access-wskjk") pod "479ebf74-62ea-4516-8c29-8d9cdf128fb4" (UID: "479ebf74-62ea-4516-8c29-8d9cdf128fb4"). InnerVolumeSpecName "kube-api-access-wskjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.840433 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/479ebf74-62ea-4516-8c29-8d9cdf128fb4-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.840502 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wskjk\" (UniqueName: \"kubernetes.io/projected/479ebf74-62ea-4516-8c29-8d9cdf128fb4-kube-api-access-wskjk\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.840515 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:27 crc kubenswrapper[4810]: I0930 08:15:27.840528 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/479ebf74-62ea-4516-8c29-8d9cdf128fb4-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.035689 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" event={"ID":"9d49935b-5628-43f0-b527-18d3b55aaa3a","Type":"ContainerStarted","Data":"f1a27e02eadf34090df67698caceb52a09bb0b432ed51107ca0f219aa8d5a9cf"} Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.037645 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" event={"ID":"9710139d-e7f3-4921-b84a-4b6525891f99","Type":"ContainerStarted","Data":"87a4d3206d3533c0be2feed46cc6606158ceb9bccc14a4337c0a0de2dd55ccc1"} Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.039343 4810 generic.go:334] "Generic (PLEG): container finished" podID="479ebf74-62ea-4516-8c29-8d9cdf128fb4" containerID="2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e" exitCode=0 Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.039399 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" event={"ID":"479ebf74-62ea-4516-8c29-8d9cdf128fb4","Type":"ContainerDied","Data":"2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e"} Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.039412 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.039418 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf" event={"ID":"479ebf74-62ea-4516-8c29-8d9cdf128fb4","Type":"ContainerDied","Data":"2f08d9d1ab22304f057a606b0e51d8555b94d249423e33ec48c1733788c6ef26"} Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.039441 4810 scope.go:117] "RemoveContainer" containerID="2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.041414 4810 generic.go:334] "Generic (PLEG): container finished" podID="80a2e9d7-a1ec-436d-8634-89c763644dec" containerID="638c48838c29dccb16889232e47ee51085f0be306b2c6857be7f7195f2fa6135" exitCode=0 Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.041448 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" event={"ID":"80a2e9d7-a1ec-436d-8634-89c763644dec","Type":"ContainerDied","Data":"638c48838c29dccb16889232e47ee51085f0be306b2c6857be7f7195f2fa6135"} Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.065654 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf"] Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.070944 4810 scope.go:117] "RemoveContainer" containerID="2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e" Sep 30 08:15:28 crc kubenswrapper[4810]: E0930 08:15:28.071387 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e\": container with ID starting with 2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e not found: ID does not exist" containerID="2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.071419 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e"} err="failed to get container status \"2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e\": rpc error: code = NotFound desc = could not find container \"2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e\": container with ID starting with 2086218772fe27ead63b8317ea654e21aff731f6348112550b02f2671287f71e not found: ID does not exist" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.071660 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xggpf"] Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.208279 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw"] Sep 30 08:15:28 crc kubenswrapper[4810]: E0930 08:15:28.208553 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="479ebf74-62ea-4516-8c29-8d9cdf128fb4" containerName="route-controller-manager" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.208567 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="479ebf74-62ea-4516-8c29-8d9cdf128fb4" containerName="route-controller-manager" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.208688 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="479ebf74-62ea-4516-8c29-8d9cdf128fb4" containerName="route-controller-manager" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.209175 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.212619 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.212911 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.213205 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.213633 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.213665 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.213879 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.239211 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw"] Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.245489 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67532a13-9a15-4089-9c33-13bf450186c2-client-ca\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.245530 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67532a13-9a15-4089-9c33-13bf450186c2-serving-cert\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.245579 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpdvg\" (UniqueName: \"kubernetes.io/projected/67532a13-9a15-4089-9c33-13bf450186c2-kube-api-access-dpdvg\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.245754 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67532a13-9a15-4089-9c33-13bf450186c2-config\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.346730 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpdvg\" (UniqueName: \"kubernetes.io/projected/67532a13-9a15-4089-9c33-13bf450186c2-kube-api-access-dpdvg\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.346789 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67532a13-9a15-4089-9c33-13bf450186c2-config\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.346899 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67532a13-9a15-4089-9c33-13bf450186c2-client-ca\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.346920 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67532a13-9a15-4089-9c33-13bf450186c2-serving-cert\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.347988 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67532a13-9a15-4089-9c33-13bf450186c2-client-ca\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.352929 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67532a13-9a15-4089-9c33-13bf450186c2-serving-cert\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.362564 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67532a13-9a15-4089-9c33-13bf450186c2-config\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.375285 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpdvg\" (UniqueName: \"kubernetes.io/projected/67532a13-9a15-4089-9c33-13bf450186c2-kube-api-access-dpdvg\") pod \"route-controller-manager-84dd7fbfb4-9zlvw\" (UID: \"67532a13-9a15-4089-9c33-13bf450186c2\") " pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.523893 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.680748 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.751094 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-config\") pod \"80a2e9d7-a1ec-436d-8634-89c763644dec\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.751182 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-client-ca\") pod \"80a2e9d7-a1ec-436d-8634-89c763644dec\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.751223 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-proxy-ca-bundles\") pod \"80a2e9d7-a1ec-436d-8634-89c763644dec\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.751301 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzj7l\" (UniqueName: \"kubernetes.io/projected/80a2e9d7-a1ec-436d-8634-89c763644dec-kube-api-access-gzj7l\") pod \"80a2e9d7-a1ec-436d-8634-89c763644dec\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.751334 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a2e9d7-a1ec-436d-8634-89c763644dec-serving-cert\") pod \"80a2e9d7-a1ec-436d-8634-89c763644dec\" (UID: \"80a2e9d7-a1ec-436d-8634-89c763644dec\") " Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.752146 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-client-ca" (OuterVolumeSpecName: "client-ca") pod "80a2e9d7-a1ec-436d-8634-89c763644dec" (UID: "80a2e9d7-a1ec-436d-8634-89c763644dec"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.752153 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "80a2e9d7-a1ec-436d-8634-89c763644dec" (UID: "80a2e9d7-a1ec-436d-8634-89c763644dec"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.752314 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-config" (OuterVolumeSpecName: "config") pod "80a2e9d7-a1ec-436d-8634-89c763644dec" (UID: "80a2e9d7-a1ec-436d-8634-89c763644dec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.766876 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80a2e9d7-a1ec-436d-8634-89c763644dec-kube-api-access-gzj7l" (OuterVolumeSpecName: "kube-api-access-gzj7l") pod "80a2e9d7-a1ec-436d-8634-89c763644dec" (UID: "80a2e9d7-a1ec-436d-8634-89c763644dec"). InnerVolumeSpecName "kube-api-access-gzj7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.767517 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80a2e9d7-a1ec-436d-8634-89c763644dec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "80a2e9d7-a1ec-436d-8634-89c763644dec" (UID: "80a2e9d7-a1ec-436d-8634-89c763644dec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.790760 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw"] Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.852686 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.852742 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.852755 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/80a2e9d7-a1ec-436d-8634-89c763644dec-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.852767 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a2e9d7-a1ec-436d-8634-89c763644dec-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:28 crc kubenswrapper[4810]: I0930 08:15:28.852779 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzj7l\" (UniqueName: \"kubernetes.io/projected/80a2e9d7-a1ec-436d-8634-89c763644dec-kube-api-access-gzj7l\") on node \"crc\" DevicePath \"\"" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.057695 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.057685 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qdcml" event={"ID":"80a2e9d7-a1ec-436d-8634-89c763644dec","Type":"ContainerDied","Data":"21c1fbdeff2ead362912ec455538dc597a1558b6220b8195e0267d21dab955b3"} Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.057805 4810 scope.go:117] "RemoveContainer" containerID="638c48838c29dccb16889232e47ee51085f0be306b2c6857be7f7195f2fa6135" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.060567 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" event={"ID":"67532a13-9a15-4089-9c33-13bf450186c2","Type":"ContainerStarted","Data":"ce872578f066afc9b09786536c560b6ecc74c5ea8a23bf767d9a13505ba2235e"} Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.063419 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.063480 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.101123 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" podStartSLOduration=2.622940694 podStartE2EDuration="19.101099357s" podCreationTimestamp="2025-09-30 08:15:10 +0000 UTC" firstStartedPulling="2025-09-30 08:15:10.734179394 +0000 UTC m=+734.186378661" lastFinishedPulling="2025-09-30 08:15:27.212338057 +0000 UTC m=+750.664537324" observedRunningTime="2025-09-30 08:15:29.097103129 +0000 UTC m=+752.549302396" watchObservedRunningTime="2025-09-30 08:15:29.101099357 +0000 UTC m=+752.553298624" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.144750 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" podStartSLOduration=3.240600446 podStartE2EDuration="19.144703613s" podCreationTimestamp="2025-09-30 08:15:10 +0000 UTC" firstStartedPulling="2025-09-30 08:15:11.327064055 +0000 UTC m=+734.779263322" lastFinishedPulling="2025-09-30 08:15:27.231167232 +0000 UTC m=+750.683366489" observedRunningTime="2025-09-30 08:15:29.126904388 +0000 UTC m=+752.579103655" watchObservedRunningTime="2025-09-30 08:15:29.144703613 +0000 UTC m=+752.596902980" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.149687 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qdcml"] Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.156977 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qdcml"] Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.320774 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="479ebf74-62ea-4516-8c29-8d9cdf128fb4" path="/var/lib/kubelet/pods/479ebf74-62ea-4516-8c29-8d9cdf128fb4/volumes" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.323722 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80a2e9d7-a1ec-436d-8634-89c763644dec" path="/var/lib/kubelet/pods/80a2e9d7-a1ec-436d-8634-89c763644dec/volumes" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.534078 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-569bc49944-7cnjw"] Sep 30 08:15:29 crc kubenswrapper[4810]: E0930 08:15:29.534319 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80a2e9d7-a1ec-436d-8634-89c763644dec" containerName="controller-manager" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.534333 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="80a2e9d7-a1ec-436d-8634-89c763644dec" containerName="controller-manager" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.535029 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="80a2e9d7-a1ec-436d-8634-89c763644dec" containerName="controller-manager" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.535460 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.537621 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.539882 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.540108 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.541543 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.541619 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.546964 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.552832 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.555571 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-569bc49944-7cnjw"] Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.559774 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2caa96df-ae5c-420a-b746-b0c8d76a986e-proxy-ca-bundles\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.559831 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2caa96df-ae5c-420a-b746-b0c8d76a986e-config\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.559883 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2caa96df-ae5c-420a-b746-b0c8d76a986e-serving-cert\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.559916 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtmkr\" (UniqueName: \"kubernetes.io/projected/2caa96df-ae5c-420a-b746-b0c8d76a986e-kube-api-access-rtmkr\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.559961 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2caa96df-ae5c-420a-b746-b0c8d76a986e-client-ca\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.660819 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2caa96df-ae5c-420a-b746-b0c8d76a986e-proxy-ca-bundles\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.660881 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2caa96df-ae5c-420a-b746-b0c8d76a986e-config\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.660923 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2caa96df-ae5c-420a-b746-b0c8d76a986e-serving-cert\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.660944 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtmkr\" (UniqueName: \"kubernetes.io/projected/2caa96df-ae5c-420a-b746-b0c8d76a986e-kube-api-access-rtmkr\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.660966 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2caa96df-ae5c-420a-b746-b0c8d76a986e-client-ca\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.661979 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2caa96df-ae5c-420a-b746-b0c8d76a986e-client-ca\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.662726 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2caa96df-ae5c-420a-b746-b0c8d76a986e-proxy-ca-bundles\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.663075 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2caa96df-ae5c-420a-b746-b0c8d76a986e-config\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.667701 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2caa96df-ae5c-420a-b746-b0c8d76a986e-serving-cert\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.720664 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtmkr\" (UniqueName: \"kubernetes.io/projected/2caa96df-ae5c-420a-b746-b0c8d76a986e-kube-api-access-rtmkr\") pod \"controller-manager-569bc49944-7cnjw\" (UID: \"2caa96df-ae5c-420a-b746-b0c8d76a986e\") " pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:29 crc kubenswrapper[4810]: I0930 08:15:29.855186 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:30 crc kubenswrapper[4810]: I0930 08:15:30.095812 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" event={"ID":"67532a13-9a15-4089-9c33-13bf450186c2","Type":"ContainerStarted","Data":"88c35b9e789b4943e21df8fbee3e8e4aafbda1b47aab089009949376248c5e8c"} Sep 30 08:15:30 crc kubenswrapper[4810]: I0930 08:15:30.098493 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:30 crc kubenswrapper[4810]: I0930 08:15:30.121395 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" podStartSLOduration=2.121337619 podStartE2EDuration="2.121337619s" podCreationTimestamp="2025-09-30 08:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:15:30.116891838 +0000 UTC m=+753.569091115" watchObservedRunningTime="2025-09-30 08:15:30.121337619 +0000 UTC m=+753.573536886" Sep 30 08:15:30 crc kubenswrapper[4810]: I0930 08:15:30.197325 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84dd7fbfb4-9zlvw" Sep 30 08:15:30 crc kubenswrapper[4810]: I0930 08:15:30.341841 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-569bc49944-7cnjw"] Sep 30 08:15:31 crc kubenswrapper[4810]: I0930 08:15:31.105332 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" event={"ID":"2caa96df-ae5c-420a-b746-b0c8d76a986e","Type":"ContainerStarted","Data":"8917dba3e34691982a2c088106ff4e2cf51669447037a2c931aa1b68cf96d32e"} Sep 30 08:15:31 crc kubenswrapper[4810]: I0930 08:15:31.105731 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" event={"ID":"2caa96df-ae5c-420a-b746-b0c8d76a986e","Type":"ContainerStarted","Data":"0f161cf07cb48d0ac5e3a5d76d198c8c475f1d8ac108189d14646777de66b6bc"} Sep 30 08:15:31 crc kubenswrapper[4810]: I0930 08:15:31.132782 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" podStartSLOduration=3.13276485 podStartE2EDuration="3.13276485s" podCreationTimestamp="2025-09-30 08:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:15:31.130196394 +0000 UTC m=+754.582395661" watchObservedRunningTime="2025-09-30 08:15:31.13276485 +0000 UTC m=+754.584964117" Sep 30 08:15:32 crc kubenswrapper[4810]: I0930 08:15:32.111714 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:32 crc kubenswrapper[4810]: I0930 08:15:32.116545 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-569bc49944-7cnjw" Sep 30 08:15:33 crc kubenswrapper[4810]: I0930 08:15:33.297794 4810 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.422741 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vghbp"] Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.425052 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.444180 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vghbp"] Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.463457 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-catalog-content\") pod \"community-operators-vghbp\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.463528 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-utilities\") pod \"community-operators-vghbp\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.463586 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxpvj\" (UniqueName: \"kubernetes.io/projected/40f0c074-b5f2-464e-954d-44a022a65e85-kube-api-access-nxpvj\") pod \"community-operators-vghbp\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.565706 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxpvj\" (UniqueName: \"kubernetes.io/projected/40f0c074-b5f2-464e-954d-44a022a65e85-kube-api-access-nxpvj\") pod \"community-operators-vghbp\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.566100 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-catalog-content\") pod \"community-operators-vghbp\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.566223 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-utilities\") pod \"community-operators-vghbp\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.566945 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-utilities\") pod \"community-operators-vghbp\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.567646 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-catalog-content\") pod \"community-operators-vghbp\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.602647 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxpvj\" (UniqueName: \"kubernetes.io/projected/40f0c074-b5f2-464e-954d-44a022a65e85-kube-api-access-nxpvj\") pod \"community-operators-vghbp\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:39 crc kubenswrapper[4810]: I0930 08:15:39.743509 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:40 crc kubenswrapper[4810]: I0930 08:15:40.346430 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vghbp"] Sep 30 08:15:40 crc kubenswrapper[4810]: I0930 08:15:40.797669 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-768c8b4c95-d2v27" Sep 30 08:15:41 crc kubenswrapper[4810]: I0930 08:15:41.174082 4810 generic.go:334] "Generic (PLEG): container finished" podID="40f0c074-b5f2-464e-954d-44a022a65e85" containerID="621337d2859f7464911701ac0d9b9c248eafa1cced0e9f97a8c4acbc0fd1c4b3" exitCode=0 Sep 30 08:15:41 crc kubenswrapper[4810]: I0930 08:15:41.174240 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vghbp" event={"ID":"40f0c074-b5f2-464e-954d-44a022a65e85","Type":"ContainerDied","Data":"621337d2859f7464911701ac0d9b9c248eafa1cced0e9f97a8c4acbc0fd1c4b3"} Sep 30 08:15:41 crc kubenswrapper[4810]: I0930 08:15:41.174509 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vghbp" event={"ID":"40f0c074-b5f2-464e-954d-44a022a65e85","Type":"ContainerStarted","Data":"2f155ce33bb436da7907333eb2d8008a34f10d20b6f960b44781501ea4aa0534"} Sep 30 08:15:46 crc kubenswrapper[4810]: I0930 08:15:46.213301 4810 generic.go:334] "Generic (PLEG): container finished" podID="40f0c074-b5f2-464e-954d-44a022a65e85" containerID="e044a68908bdefd33ef52ea9de54fde34996db6d352d128039e3c1c10b07092e" exitCode=0 Sep 30 08:15:46 crc kubenswrapper[4810]: I0930 08:15:46.213432 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vghbp" event={"ID":"40f0c074-b5f2-464e-954d-44a022a65e85","Type":"ContainerDied","Data":"e044a68908bdefd33ef52ea9de54fde34996db6d352d128039e3c1c10b07092e"} Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.219683 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rbdnf"] Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.221084 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.241118 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbdnf"] Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.260604 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-utilities\") pod \"redhat-marketplace-rbdnf\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.260665 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skgh5\" (UniqueName: \"kubernetes.io/projected/1b6c11bf-a794-4afb-b023-40279fa07fa4-kube-api-access-skgh5\") pod \"redhat-marketplace-rbdnf\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.260708 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-catalog-content\") pod \"redhat-marketplace-rbdnf\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.361787 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-utilities\") pod \"redhat-marketplace-rbdnf\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.361848 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skgh5\" (UniqueName: \"kubernetes.io/projected/1b6c11bf-a794-4afb-b023-40279fa07fa4-kube-api-access-skgh5\") pod \"redhat-marketplace-rbdnf\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.361899 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-catalog-content\") pod \"redhat-marketplace-rbdnf\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.362454 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-catalog-content\") pod \"redhat-marketplace-rbdnf\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.363129 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-utilities\") pod \"redhat-marketplace-rbdnf\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.389635 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skgh5\" (UniqueName: \"kubernetes.io/projected/1b6c11bf-a794-4afb-b023-40279fa07fa4-kube-api-access-skgh5\") pod \"redhat-marketplace-rbdnf\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:47 crc kubenswrapper[4810]: I0930 08:15:47.538342 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:48 crc kubenswrapper[4810]: W0930 08:15:48.084580 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b6c11bf_a794_4afb_b023_40279fa07fa4.slice/crio-390b40355e5abeee36d4d41791566ca4034e791f80256eefc97a4cca57eaedeb WatchSource:0}: Error finding container 390b40355e5abeee36d4d41791566ca4034e791f80256eefc97a4cca57eaedeb: Status 404 returned error can't find the container with id 390b40355e5abeee36d4d41791566ca4034e791f80256eefc97a4cca57eaedeb Sep 30 08:15:48 crc kubenswrapper[4810]: I0930 08:15:48.089545 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbdnf"] Sep 30 08:15:48 crc kubenswrapper[4810]: I0930 08:15:48.228302 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbdnf" event={"ID":"1b6c11bf-a794-4afb-b023-40279fa07fa4","Type":"ContainerStarted","Data":"390b40355e5abeee36d4d41791566ca4034e791f80256eefc97a4cca57eaedeb"} Sep 30 08:15:52 crc kubenswrapper[4810]: I0930 08:15:52.260770 4810 generic.go:334] "Generic (PLEG): container finished" podID="1b6c11bf-a794-4afb-b023-40279fa07fa4" containerID="d0b43d2bb3af031604f761151334a3716bc8ebc423dd9ff052190f7770f89f12" exitCode=0 Sep 30 08:15:52 crc kubenswrapper[4810]: I0930 08:15:52.260848 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbdnf" event={"ID":"1b6c11bf-a794-4afb-b023-40279fa07fa4","Type":"ContainerDied","Data":"d0b43d2bb3af031604f761151334a3716bc8ebc423dd9ff052190f7770f89f12"} Sep 30 08:15:52 crc kubenswrapper[4810]: I0930 08:15:52.271172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vghbp" event={"ID":"40f0c074-b5f2-464e-954d-44a022a65e85","Type":"ContainerStarted","Data":"fd86e421fda1de21ec5fa21d0ab832bbfd1df8c7893c4bae4eb2520a72a9b2b4"} Sep 30 08:15:52 crc kubenswrapper[4810]: I0930 08:15:52.298461 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vghbp" podStartSLOduration=2.455835755 podStartE2EDuration="13.298440727s" podCreationTimestamp="2025-09-30 08:15:39 +0000 UTC" firstStartedPulling="2025-09-30 08:15:41.176748676 +0000 UTC m=+764.628947943" lastFinishedPulling="2025-09-30 08:15:52.019353608 +0000 UTC m=+775.471552915" observedRunningTime="2025-09-30 08:15:52.296278113 +0000 UTC m=+775.748477390" watchObservedRunningTime="2025-09-30 08:15:52.298440727 +0000 UTC m=+775.750639994" Sep 30 08:15:53 crc kubenswrapper[4810]: I0930 08:15:53.279213 4810 generic.go:334] "Generic (PLEG): container finished" podID="1b6c11bf-a794-4afb-b023-40279fa07fa4" containerID="6598e436ba90115eb39dffaa6616aa4970ca2ea347d2e146274ac3babe98b668" exitCode=0 Sep 30 08:15:53 crc kubenswrapper[4810]: I0930 08:15:53.279314 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbdnf" event={"ID":"1b6c11bf-a794-4afb-b023-40279fa07fa4","Type":"ContainerDied","Data":"6598e436ba90115eb39dffaa6616aa4970ca2ea347d2e146274ac3babe98b668"} Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.290388 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbdnf" event={"ID":"1b6c11bf-a794-4afb-b023-40279fa07fa4","Type":"ContainerStarted","Data":"86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d"} Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.319392 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rbdnf" podStartSLOduration=5.87512315 podStartE2EDuration="7.319368624s" podCreationTimestamp="2025-09-30 08:15:47 +0000 UTC" firstStartedPulling="2025-09-30 08:15:52.26295092 +0000 UTC m=+775.715150187" lastFinishedPulling="2025-09-30 08:15:53.707196384 +0000 UTC m=+777.159395661" observedRunningTime="2025-09-30 08:15:54.313512551 +0000 UTC m=+777.765711848" watchObservedRunningTime="2025-09-30 08:15:54.319368624 +0000 UTC m=+777.771567911" Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.508724 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fsz7d"] Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.511015 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.527137 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fsz7d"] Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.606309 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-utilities\") pod \"certified-operators-fsz7d\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.606406 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-catalog-content\") pod \"certified-operators-fsz7d\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.606434 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5c4g\" (UniqueName: \"kubernetes.io/projected/b73b1482-e35d-4784-8564-80f798b39727-kube-api-access-t5c4g\") pod \"certified-operators-fsz7d\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.707896 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-utilities\") pod \"certified-operators-fsz7d\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.707963 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-catalog-content\") pod \"certified-operators-fsz7d\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.707986 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5c4g\" (UniqueName: \"kubernetes.io/projected/b73b1482-e35d-4784-8564-80f798b39727-kube-api-access-t5c4g\") pod \"certified-operators-fsz7d\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.708630 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-utilities\") pod \"certified-operators-fsz7d\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.708663 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-catalog-content\") pod \"certified-operators-fsz7d\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.729414 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5c4g\" (UniqueName: \"kubernetes.io/projected/b73b1482-e35d-4784-8564-80f798b39727-kube-api-access-t5c4g\") pod \"certified-operators-fsz7d\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:15:54 crc kubenswrapper[4810]: I0930 08:15:54.827894 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:15:55 crc kubenswrapper[4810]: I0930 08:15:55.373123 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fsz7d"] Sep 30 08:15:55 crc kubenswrapper[4810]: W0930 08:15:55.375504 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb73b1482_e35d_4784_8564_80f798b39727.slice/crio-41e99b8218e4efbcfcecf9a0277b05aafe3e48036108b648ef908da59f630bd5 WatchSource:0}: Error finding container 41e99b8218e4efbcfcecf9a0277b05aafe3e48036108b648ef908da59f630bd5: Status 404 returned error can't find the container with id 41e99b8218e4efbcfcecf9a0277b05aafe3e48036108b648ef908da59f630bd5 Sep 30 08:15:56 crc kubenswrapper[4810]: I0930 08:15:56.305665 4810 generic.go:334] "Generic (PLEG): container finished" podID="b73b1482-e35d-4784-8564-80f798b39727" containerID="7b181134fb76dbedcc0f3cb8069ccffdb3d85356337ea37cb688ed4fe2076f56" exitCode=0 Sep 30 08:15:56 crc kubenswrapper[4810]: I0930 08:15:56.305759 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsz7d" event={"ID":"b73b1482-e35d-4784-8564-80f798b39727","Type":"ContainerDied","Data":"7b181134fb76dbedcc0f3cb8069ccffdb3d85356337ea37cb688ed4fe2076f56"} Sep 30 08:15:56 crc kubenswrapper[4810]: I0930 08:15:56.305805 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsz7d" event={"ID":"b73b1482-e35d-4784-8564-80f798b39727","Type":"ContainerStarted","Data":"41e99b8218e4efbcfcecf9a0277b05aafe3e48036108b648ef908da59f630bd5"} Sep 30 08:15:57 crc kubenswrapper[4810]: I0930 08:15:57.539414 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:57 crc kubenswrapper[4810]: I0930 08:15:57.540518 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:57 crc kubenswrapper[4810]: I0930 08:15:57.607459 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:58 crc kubenswrapper[4810]: I0930 08:15:58.330591 4810 generic.go:334] "Generic (PLEG): container finished" podID="b73b1482-e35d-4784-8564-80f798b39727" containerID="641dfed407850c9a2ded6c2eb3c7419260c5f8060e829ffe0f2541746ed9257c" exitCode=0 Sep 30 08:15:58 crc kubenswrapper[4810]: I0930 08:15:58.330818 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsz7d" event={"ID":"b73b1482-e35d-4784-8564-80f798b39727","Type":"ContainerDied","Data":"641dfed407850c9a2ded6c2eb3c7419260c5f8060e829ffe0f2541746ed9257c"} Sep 30 08:15:58 crc kubenswrapper[4810]: I0930 08:15:58.398136 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:15:58 crc kubenswrapper[4810]: I0930 08:15:58.894395 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbdnf"] Sep 30 08:15:59 crc kubenswrapper[4810]: I0930 08:15:59.744639 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:59 crc kubenswrapper[4810]: I0930 08:15:59.744940 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:15:59 crc kubenswrapper[4810]: I0930 08:15:59.811891 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:16:00 crc kubenswrapper[4810]: I0930 08:16:00.352155 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsz7d" event={"ID":"b73b1482-e35d-4784-8564-80f798b39727","Type":"ContainerStarted","Data":"fbe57a2f15bb0a2c9ca27ff4f8ae1af1557ceae18a755249c091536132797686"} Sep 30 08:16:00 crc kubenswrapper[4810]: I0930 08:16:00.352874 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rbdnf" podUID="1b6c11bf-a794-4afb-b023-40279fa07fa4" containerName="registry-server" containerID="cri-o://86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d" gracePeriod=2 Sep 30 08:16:00 crc kubenswrapper[4810]: I0930 08:16:00.414322 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:16:00 crc kubenswrapper[4810]: I0930 08:16:00.432135 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fsz7d" podStartSLOduration=3.500478408 podStartE2EDuration="6.432118246s" podCreationTimestamp="2025-09-30 08:15:54 +0000 UTC" firstStartedPulling="2025-09-30 08:15:56.30774062 +0000 UTC m=+779.759939897" lastFinishedPulling="2025-09-30 08:15:59.239380458 +0000 UTC m=+782.691579735" observedRunningTime="2025-09-30 08:16:00.390589372 +0000 UTC m=+783.842788659" watchObservedRunningTime="2025-09-30 08:16:00.432118246 +0000 UTC m=+783.884317513" Sep 30 08:16:00 crc kubenswrapper[4810]: I0930 08:16:00.463323 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-74b66975c5-rtrd5" Sep 30 08:16:00 crc kubenswrapper[4810]: I0930 08:16:00.939754 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.015956 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-utilities\") pod \"1b6c11bf-a794-4afb-b023-40279fa07fa4\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.016024 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-catalog-content\") pod \"1b6c11bf-a794-4afb-b023-40279fa07fa4\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.016091 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skgh5\" (UniqueName: \"kubernetes.io/projected/1b6c11bf-a794-4afb-b023-40279fa07fa4-kube-api-access-skgh5\") pod \"1b6c11bf-a794-4afb-b023-40279fa07fa4\" (UID: \"1b6c11bf-a794-4afb-b023-40279fa07fa4\") " Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.017134 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-utilities" (OuterVolumeSpecName: "utilities") pod "1b6c11bf-a794-4afb-b023-40279fa07fa4" (UID: "1b6c11bf-a794-4afb-b023-40279fa07fa4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.028200 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b6c11bf-a794-4afb-b023-40279fa07fa4" (UID: "1b6c11bf-a794-4afb-b023-40279fa07fa4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.029636 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b6c11bf-a794-4afb-b023-40279fa07fa4-kube-api-access-skgh5" (OuterVolumeSpecName: "kube-api-access-skgh5") pod "1b6c11bf-a794-4afb-b023-40279fa07fa4" (UID: "1b6c11bf-a794-4afb-b023-40279fa07fa4"). InnerVolumeSpecName "kube-api-access-skgh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.118203 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.118244 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b6c11bf-a794-4afb-b023-40279fa07fa4-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.118257 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skgh5\" (UniqueName: \"kubernetes.io/projected/1b6c11bf-a794-4afb-b023-40279fa07fa4-kube-api-access-skgh5\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.267916 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-wfqtl"] Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.268235 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b6c11bf-a794-4afb-b023-40279fa07fa4" containerName="extract-utilities" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.268250 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b6c11bf-a794-4afb-b023-40279fa07fa4" containerName="extract-utilities" Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.268293 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b6c11bf-a794-4afb-b023-40279fa07fa4" containerName="extract-content" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.268303 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b6c11bf-a794-4afb-b023-40279fa07fa4" containerName="extract-content" Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.268315 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b6c11bf-a794-4afb-b023-40279fa07fa4" containerName="registry-server" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.268324 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b6c11bf-a794-4afb-b023-40279fa07fa4" containerName="registry-server" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.268455 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b6c11bf-a794-4afb-b023-40279fa07fa4" containerName="registry-server" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.270580 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.273847 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.275004 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6fzvm" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.291893 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4"] Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.292831 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.293693 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.294307 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.316431 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4"] Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.322722 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-frr-sockets\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.322778 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46gzg\" (UniqueName: \"kubernetes.io/projected/4debe3c7-7058-4f37-863c-0c0e57818c86-kube-api-access-46gzg\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.322829 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4debe3c7-7058-4f37-863c-0c0e57818c86-frr-startup\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.322857 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-frr-conf\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.322879 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-metrics\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.322896 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4debe3c7-7058-4f37-863c-0c0e57818c86-metrics-certs\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.322914 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-reloader\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.362781 4810 generic.go:334] "Generic (PLEG): container finished" podID="1b6c11bf-a794-4afb-b023-40279fa07fa4" containerID="86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d" exitCode=0 Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.363702 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbdnf" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.364182 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbdnf" event={"ID":"1b6c11bf-a794-4afb-b023-40279fa07fa4","Type":"ContainerDied","Data":"86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d"} Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.364220 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbdnf" event={"ID":"1b6c11bf-a794-4afb-b023-40279fa07fa4","Type":"ContainerDied","Data":"390b40355e5abeee36d4d41791566ca4034e791f80256eefc97a4cca57eaedeb"} Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.364245 4810 scope.go:117] "RemoveContainer" containerID="86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.374813 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-gjfhs"] Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.375995 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-gjfhs" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.382949 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.383493 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.383707 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-hqkcm" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.385939 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.390368 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-qr7xw"] Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.391560 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.399008 4810 scope.go:117] "RemoveContainer" containerID="6598e436ba90115eb39dffaa6616aa4970ca2ea347d2e146274ac3babe98b668" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.399202 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.403122 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbdnf"] Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.415755 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbdnf"] Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.424785 4810 scope.go:117] "RemoveContainer" containerID="d0b43d2bb3af031604f761151334a3716bc8ebc423dd9ff052190f7770f89f12" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.424973 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-qr7xw"] Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425568 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4debe3c7-7058-4f37-863c-0c0e57818c86-metrics-certs\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425599 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b5cx\" (UniqueName: \"kubernetes.io/projected/2f29bb1f-4702-4fa8-8a9e-692047b356f9-kube-api-access-4b5cx\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425620 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-reloader\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425643 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-frr-sockets\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425668 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4gq5\" (UniqueName: \"kubernetes.io/projected/ad0bc482-465c-433b-bd10-82ed2f79b6ad-kube-api-access-p4gq5\") pod \"frr-k8s-webhook-server-5478bdb765-2cfm4\" (UID: \"ad0bc482-465c-433b-bd10-82ed2f79b6ad\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425684 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ad0bc482-465c-433b-bd10-82ed2f79b6ad-cert\") pod \"frr-k8s-webhook-server-5478bdb765-2cfm4\" (UID: \"ad0bc482-465c-433b-bd10-82ed2f79b6ad\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425713 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46gzg\" (UniqueName: \"kubernetes.io/projected/4debe3c7-7058-4f37-863c-0c0e57818c86-kube-api-access-46gzg\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425734 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2f29bb1f-4702-4fa8-8a9e-692047b356f9-metallb-excludel2\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425762 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-memberlist\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425782 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4debe3c7-7058-4f37-863c-0c0e57818c86-frr-startup\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425817 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-frr-conf\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425835 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-metrics-certs\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.425854 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-metrics\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.426377 4810 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.426418 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4debe3c7-7058-4f37-863c-0c0e57818c86-metrics-certs podName:4debe3c7-7058-4f37-863c-0c0e57818c86 nodeName:}" failed. No retries permitted until 2025-09-30 08:16:01.926404483 +0000 UTC m=+785.378603750 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4debe3c7-7058-4f37-863c-0c0e57818c86-metrics-certs") pod "frr-k8s-wfqtl" (UID: "4debe3c7-7058-4f37-863c-0c0e57818c86") : secret "frr-k8s-certs-secret" not found Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.427243 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4debe3c7-7058-4f37-863c-0c0e57818c86-frr-startup\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.427679 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-frr-conf\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.427857 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-metrics\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.429146 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-frr-sockets\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.429931 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4debe3c7-7058-4f37-863c-0c0e57818c86-reloader\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.449596 4810 scope.go:117] "RemoveContainer" containerID="86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d" Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.450084 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d\": container with ID starting with 86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d not found: ID does not exist" containerID="86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.450112 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d"} err="failed to get container status \"86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d\": rpc error: code = NotFound desc = could not find container \"86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d\": container with ID starting with 86f771fe3d56afe5a5d547430f8c254468d7dd29f792d95f72f921de2aed707d not found: ID does not exist" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.450134 4810 scope.go:117] "RemoveContainer" containerID="6598e436ba90115eb39dffaa6616aa4970ca2ea347d2e146274ac3babe98b668" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.450423 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46gzg\" (UniqueName: \"kubernetes.io/projected/4debe3c7-7058-4f37-863c-0c0e57818c86-kube-api-access-46gzg\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.450422 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6598e436ba90115eb39dffaa6616aa4970ca2ea347d2e146274ac3babe98b668\": container with ID starting with 6598e436ba90115eb39dffaa6616aa4970ca2ea347d2e146274ac3babe98b668 not found: ID does not exist" containerID="6598e436ba90115eb39dffaa6616aa4970ca2ea347d2e146274ac3babe98b668" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.450478 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6598e436ba90115eb39dffaa6616aa4970ca2ea347d2e146274ac3babe98b668"} err="failed to get container status \"6598e436ba90115eb39dffaa6616aa4970ca2ea347d2e146274ac3babe98b668\": rpc error: code = NotFound desc = could not find container \"6598e436ba90115eb39dffaa6616aa4970ca2ea347d2e146274ac3babe98b668\": container with ID starting with 6598e436ba90115eb39dffaa6616aa4970ca2ea347d2e146274ac3babe98b668 not found: ID does not exist" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.450506 4810 scope.go:117] "RemoveContainer" containerID="d0b43d2bb3af031604f761151334a3716bc8ebc423dd9ff052190f7770f89f12" Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.450775 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0b43d2bb3af031604f761151334a3716bc8ebc423dd9ff052190f7770f89f12\": container with ID starting with d0b43d2bb3af031604f761151334a3716bc8ebc423dd9ff052190f7770f89f12 not found: ID does not exist" containerID="d0b43d2bb3af031604f761151334a3716bc8ebc423dd9ff052190f7770f89f12" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.450797 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0b43d2bb3af031604f761151334a3716bc8ebc423dd9ff052190f7770f89f12"} err="failed to get container status \"d0b43d2bb3af031604f761151334a3716bc8ebc423dd9ff052190f7770f89f12\": rpc error: code = NotFound desc = could not find container \"d0b43d2bb3af031604f761151334a3716bc8ebc423dd9ff052190f7770f89f12\": container with ID starting with d0b43d2bb3af031604f761151334a3716bc8ebc423dd9ff052190f7770f89f12 not found: ID does not exist" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.526630 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-metrics-certs\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.526686 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b5cx\" (UniqueName: \"kubernetes.io/projected/2f29bb1f-4702-4fa8-8a9e-692047b356f9-kube-api-access-4b5cx\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.526720 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4gq5\" (UniqueName: \"kubernetes.io/projected/ad0bc482-465c-433b-bd10-82ed2f79b6ad-kube-api-access-p4gq5\") pod \"frr-k8s-webhook-server-5478bdb765-2cfm4\" (UID: \"ad0bc482-465c-433b-bd10-82ed2f79b6ad\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.526754 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ad0bc482-465c-433b-bd10-82ed2f79b6ad-cert\") pod \"frr-k8s-webhook-server-5478bdb765-2cfm4\" (UID: \"ad0bc482-465c-433b-bd10-82ed2f79b6ad\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.526776 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f9b2513-5122-4757-b48c-4f0e8c9aaf05-metrics-certs\") pod \"controller-5d688f5ffc-qr7xw\" (UID: \"4f9b2513-5122-4757-b48c-4f0e8c9aaf05\") " pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.526794 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhhb5\" (UniqueName: \"kubernetes.io/projected/4f9b2513-5122-4757-b48c-4f0e8c9aaf05-kube-api-access-lhhb5\") pod \"controller-5d688f5ffc-qr7xw\" (UID: \"4f9b2513-5122-4757-b48c-4f0e8c9aaf05\") " pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.526816 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f9b2513-5122-4757-b48c-4f0e8c9aaf05-cert\") pod \"controller-5d688f5ffc-qr7xw\" (UID: \"4f9b2513-5122-4757-b48c-4f0e8c9aaf05\") " pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.526845 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2f29bb1f-4702-4fa8-8a9e-692047b356f9-metallb-excludel2\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.526813 4810 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.526900 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-memberlist\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.526969 4810 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.526985 4810 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.527030 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-metrics-certs podName:2f29bb1f-4702-4fa8-8a9e-692047b356f9 nodeName:}" failed. No retries permitted until 2025-09-30 08:16:02.026913486 +0000 UTC m=+785.479112753 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-metrics-certs") pod "speaker-gjfhs" (UID: "2f29bb1f-4702-4fa8-8a9e-692047b356f9") : secret "speaker-certs-secret" not found Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.527044 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad0bc482-465c-433b-bd10-82ed2f79b6ad-cert podName:ad0bc482-465c-433b-bd10-82ed2f79b6ad nodeName:}" failed. No retries permitted until 2025-09-30 08:16:02.02703892 +0000 UTC m=+785.479238177 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ad0bc482-465c-433b-bd10-82ed2f79b6ad-cert") pod "frr-k8s-webhook-server-5478bdb765-2cfm4" (UID: "ad0bc482-465c-433b-bd10-82ed2f79b6ad") : secret "frr-k8s-webhook-server-cert" not found Sep 30 08:16:01 crc kubenswrapper[4810]: E0930 08:16:01.527057 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-memberlist podName:2f29bb1f-4702-4fa8-8a9e-692047b356f9 nodeName:}" failed. No retries permitted until 2025-09-30 08:16:02.02705157 +0000 UTC m=+785.479250827 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-memberlist") pod "speaker-gjfhs" (UID: "2f29bb1f-4702-4fa8-8a9e-692047b356f9") : secret "metallb-memberlist" not found Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.527474 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2f29bb1f-4702-4fa8-8a9e-692047b356f9-metallb-excludel2\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.548497 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b5cx\" (UniqueName: \"kubernetes.io/projected/2f29bb1f-4702-4fa8-8a9e-692047b356f9-kube-api-access-4b5cx\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.548516 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4gq5\" (UniqueName: \"kubernetes.io/projected/ad0bc482-465c-433b-bd10-82ed2f79b6ad-kube-api-access-p4gq5\") pod \"frr-k8s-webhook-server-5478bdb765-2cfm4\" (UID: \"ad0bc482-465c-433b-bd10-82ed2f79b6ad\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.628568 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f9b2513-5122-4757-b48c-4f0e8c9aaf05-metrics-certs\") pod \"controller-5d688f5ffc-qr7xw\" (UID: \"4f9b2513-5122-4757-b48c-4f0e8c9aaf05\") " pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.628609 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhhb5\" (UniqueName: \"kubernetes.io/projected/4f9b2513-5122-4757-b48c-4f0e8c9aaf05-kube-api-access-lhhb5\") pod \"controller-5d688f5ffc-qr7xw\" (UID: \"4f9b2513-5122-4757-b48c-4f0e8c9aaf05\") " pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.628629 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f9b2513-5122-4757-b48c-4f0e8c9aaf05-cert\") pod \"controller-5d688f5ffc-qr7xw\" (UID: \"4f9b2513-5122-4757-b48c-4f0e8c9aaf05\") " pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.630592 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.639199 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f9b2513-5122-4757-b48c-4f0e8c9aaf05-metrics-certs\") pod \"controller-5d688f5ffc-qr7xw\" (UID: \"4f9b2513-5122-4757-b48c-4f0e8c9aaf05\") " pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.642573 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f9b2513-5122-4757-b48c-4f0e8c9aaf05-cert\") pod \"controller-5d688f5ffc-qr7xw\" (UID: \"4f9b2513-5122-4757-b48c-4f0e8c9aaf05\") " pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.647114 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhhb5\" (UniqueName: \"kubernetes.io/projected/4f9b2513-5122-4757-b48c-4f0e8c9aaf05-kube-api-access-lhhb5\") pod \"controller-5d688f5ffc-qr7xw\" (UID: \"4f9b2513-5122-4757-b48c-4f0e8c9aaf05\") " pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.711633 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.935838 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4debe3c7-7058-4f37-863c-0c0e57818c86-metrics-certs\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:01 crc kubenswrapper[4810]: I0930 08:16:01.941789 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4debe3c7-7058-4f37-863c-0c0e57818c86-metrics-certs\") pod \"frr-k8s-wfqtl\" (UID: \"4debe3c7-7058-4f37-863c-0c0e57818c86\") " pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.037805 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-memberlist\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.037930 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-metrics-certs\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.037994 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ad0bc482-465c-433b-bd10-82ed2f79b6ad-cert\") pod \"frr-k8s-webhook-server-5478bdb765-2cfm4\" (UID: \"ad0bc482-465c-433b-bd10-82ed2f79b6ad\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" Sep 30 08:16:02 crc kubenswrapper[4810]: E0930 08:16:02.038007 4810 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 30 08:16:02 crc kubenswrapper[4810]: E0930 08:16:02.038115 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-memberlist podName:2f29bb1f-4702-4fa8-8a9e-692047b356f9 nodeName:}" failed. No retries permitted until 2025-09-30 08:16:03.038080058 +0000 UTC m=+786.490279365 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-memberlist") pod "speaker-gjfhs" (UID: "2f29bb1f-4702-4fa8-8a9e-692047b356f9") : secret "metallb-memberlist" not found Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.041242 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-metrics-certs\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.042366 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ad0bc482-465c-433b-bd10-82ed2f79b6ad-cert\") pod \"frr-k8s-webhook-server-5478bdb765-2cfm4\" (UID: \"ad0bc482-465c-433b-bd10-82ed2f79b6ad\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.186918 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.211055 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.257253 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-qr7xw"] Sep 30 08:16:02 crc kubenswrapper[4810]: W0930 08:16:02.264583 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f9b2513_5122_4757_b48c_4f0e8c9aaf05.slice/crio-ab70099473be16001210f9fb3594dff56eda0770b28c97bcd4bc19afe0ca970c WatchSource:0}: Error finding container ab70099473be16001210f9fb3594dff56eda0770b28c97bcd4bc19afe0ca970c: Status 404 returned error can't find the container with id ab70099473be16001210f9fb3594dff56eda0770b28c97bcd4bc19afe0ca970c Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.390616 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-qr7xw" event={"ID":"4f9b2513-5122-4757-b48c-4f0e8c9aaf05","Type":"ContainerStarted","Data":"ab70099473be16001210f9fb3594dff56eda0770b28c97bcd4bc19afe0ca970c"} Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.473456 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4"] Sep 30 08:16:02 crc kubenswrapper[4810]: W0930 08:16:02.489494 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad0bc482_465c_433b_bd10_82ed2f79b6ad.slice/crio-f8aad8b59a4e957a0879b03c60c30ea80f76e9eb801ab0443ed06a0606fe7ed7 WatchSource:0}: Error finding container f8aad8b59a4e957a0879b03c60c30ea80f76e9eb801ab0443ed06a0606fe7ed7: Status 404 returned error can't find the container with id f8aad8b59a4e957a0879b03c60c30ea80f76e9eb801ab0443ed06a0606fe7ed7 Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.694606 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vghbp"] Sep 30 08:16:02 crc kubenswrapper[4810]: I0930 08:16:02.694925 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vghbp" podUID="40f0c074-b5f2-464e-954d-44a022a65e85" containerName="registry-server" containerID="cri-o://fd86e421fda1de21ec5fa21d0ab832bbfd1df8c7893c4bae4eb2520a72a9b2b4" gracePeriod=2 Sep 30 08:16:03 crc kubenswrapper[4810]: I0930 08:16:03.055576 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-memberlist\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:03 crc kubenswrapper[4810]: I0930 08:16:03.062254 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2f29bb1f-4702-4fa8-8a9e-692047b356f9-memberlist\") pod \"speaker-gjfhs\" (UID: \"2f29bb1f-4702-4fa8-8a9e-692047b356f9\") " pod="metallb-system/speaker-gjfhs" Sep 30 08:16:03 crc kubenswrapper[4810]: I0930 08:16:03.200231 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-gjfhs" Sep 30 08:16:03 crc kubenswrapper[4810]: W0930 08:16:03.224351 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f29bb1f_4702_4fa8_8a9e_692047b356f9.slice/crio-f352175c43fcc824076023998b536f77b8e2fdda486834e882b0856c35ff5aaf WatchSource:0}: Error finding container f352175c43fcc824076023998b536f77b8e2fdda486834e882b0856c35ff5aaf: Status 404 returned error can't find the container with id f352175c43fcc824076023998b536f77b8e2fdda486834e882b0856c35ff5aaf Sep 30 08:16:03 crc kubenswrapper[4810]: I0930 08:16:03.316712 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b6c11bf-a794-4afb-b023-40279fa07fa4" path="/var/lib/kubelet/pods/1b6c11bf-a794-4afb-b023-40279fa07fa4/volumes" Sep 30 08:16:03 crc kubenswrapper[4810]: I0930 08:16:03.401680 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-gjfhs" event={"ID":"2f29bb1f-4702-4fa8-8a9e-692047b356f9","Type":"ContainerStarted","Data":"f352175c43fcc824076023998b536f77b8e2fdda486834e882b0856c35ff5aaf"} Sep 30 08:16:03 crc kubenswrapper[4810]: I0930 08:16:03.404000 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-qr7xw" event={"ID":"4f9b2513-5122-4757-b48c-4f0e8c9aaf05","Type":"ContainerStarted","Data":"9cd3ed25506034ce881fab50dad6adbaff2d002d487ee4efb6fb501d631ad55d"} Sep 30 08:16:03 crc kubenswrapper[4810]: I0930 08:16:03.404055 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-qr7xw" event={"ID":"4f9b2513-5122-4757-b48c-4f0e8c9aaf05","Type":"ContainerStarted","Data":"323098e6be203813e8364408391f7acbd1e38b7faec13857421cac8278c2eb46"} Sep 30 08:16:03 crc kubenswrapper[4810]: I0930 08:16:03.405286 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" event={"ID":"ad0bc482-465c-433b-bd10-82ed2f79b6ad","Type":"ContainerStarted","Data":"f8aad8b59a4e957a0879b03c60c30ea80f76e9eb801ab0443ed06a0606fe7ed7"} Sep 30 08:16:03 crc kubenswrapper[4810]: I0930 08:16:03.406241 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wfqtl" event={"ID":"4debe3c7-7058-4f37-863c-0c0e57818c86","Type":"ContainerStarted","Data":"6110e6612776c6f4adcc8522206b777a8ed94eca52bcc354a398d774b1a3e524"} Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.417609 4810 generic.go:334] "Generic (PLEG): container finished" podID="40f0c074-b5f2-464e-954d-44a022a65e85" containerID="fd86e421fda1de21ec5fa21d0ab832bbfd1df8c7893c4bae4eb2520a72a9b2b4" exitCode=0 Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.417712 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vghbp" event={"ID":"40f0c074-b5f2-464e-954d-44a022a65e85","Type":"ContainerDied","Data":"fd86e421fda1de21ec5fa21d0ab832bbfd1df8c7893c4bae4eb2520a72a9b2b4"} Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.421777 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-gjfhs" event={"ID":"2f29bb1f-4702-4fa8-8a9e-692047b356f9","Type":"ContainerStarted","Data":"13c022a53784d021fa240ed49795601f78160e71f5b706ca80a821fa9cfd39fd"} Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.421864 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.765893 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.793779 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-qr7xw" podStartSLOduration=3.793758618 podStartE2EDuration="3.793758618s" podCreationTimestamp="2025-09-30 08:16:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:16:04.451922359 +0000 UTC m=+787.904121626" watchObservedRunningTime="2025-09-30 08:16:04.793758618 +0000 UTC m=+788.245957905" Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.829014 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.829104 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.881126 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.896556 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-utilities\") pod \"40f0c074-b5f2-464e-954d-44a022a65e85\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.896635 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-catalog-content\") pod \"40f0c074-b5f2-464e-954d-44a022a65e85\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.896707 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxpvj\" (UniqueName: \"kubernetes.io/projected/40f0c074-b5f2-464e-954d-44a022a65e85-kube-api-access-nxpvj\") pod \"40f0c074-b5f2-464e-954d-44a022a65e85\" (UID: \"40f0c074-b5f2-464e-954d-44a022a65e85\") " Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.897524 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-utilities" (OuterVolumeSpecName: "utilities") pod "40f0c074-b5f2-464e-954d-44a022a65e85" (UID: "40f0c074-b5f2-464e-954d-44a022a65e85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.902326 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40f0c074-b5f2-464e-954d-44a022a65e85-kube-api-access-nxpvj" (OuterVolumeSpecName: "kube-api-access-nxpvj") pod "40f0c074-b5f2-464e-954d-44a022a65e85" (UID: "40f0c074-b5f2-464e-954d-44a022a65e85"). InnerVolumeSpecName "kube-api-access-nxpvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.949185 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40f0c074-b5f2-464e-954d-44a022a65e85" (UID: "40f0c074-b5f2-464e-954d-44a022a65e85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.998403 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.998436 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f0c074-b5f2-464e-954d-44a022a65e85-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:04 crc kubenswrapper[4810]: I0930 08:16:04.998450 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxpvj\" (UniqueName: \"kubernetes.io/projected/40f0c074-b5f2-464e-954d-44a022a65e85-kube-api-access-nxpvj\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:05 crc kubenswrapper[4810]: I0930 08:16:05.431939 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vghbp" event={"ID":"40f0c074-b5f2-464e-954d-44a022a65e85","Type":"ContainerDied","Data":"2f155ce33bb436da7907333eb2d8008a34f10d20b6f960b44781501ea4aa0534"} Sep 30 08:16:05 crc kubenswrapper[4810]: I0930 08:16:05.432344 4810 scope.go:117] "RemoveContainer" containerID="fd86e421fda1de21ec5fa21d0ab832bbfd1df8c7893c4bae4eb2520a72a9b2b4" Sep 30 08:16:05 crc kubenswrapper[4810]: I0930 08:16:05.432761 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vghbp" Sep 30 08:16:05 crc kubenswrapper[4810]: I0930 08:16:05.447473 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-gjfhs" event={"ID":"2f29bb1f-4702-4fa8-8a9e-692047b356f9","Type":"ContainerStarted","Data":"665b4ff171b2e297df5bf959ee2945e276d97e97c17452f6ce25f1bfcbd9b9fc"} Sep 30 08:16:05 crc kubenswrapper[4810]: I0930 08:16:05.448051 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-gjfhs" Sep 30 08:16:05 crc kubenswrapper[4810]: I0930 08:16:05.456435 4810 scope.go:117] "RemoveContainer" containerID="e044a68908bdefd33ef52ea9de54fde34996db6d352d128039e3c1c10b07092e" Sep 30 08:16:05 crc kubenswrapper[4810]: I0930 08:16:05.462850 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vghbp"] Sep 30 08:16:05 crc kubenswrapper[4810]: I0930 08:16:05.466353 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vghbp"] Sep 30 08:16:05 crc kubenswrapper[4810]: I0930 08:16:05.485059 4810 scope.go:117] "RemoveContainer" containerID="621337d2859f7464911701ac0d9b9c248eafa1cced0e9f97a8c4acbc0fd1c4b3" Sep 30 08:16:05 crc kubenswrapper[4810]: I0930 08:16:05.497060 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:16:05 crc kubenswrapper[4810]: I0930 08:16:05.618157 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-gjfhs" podStartSLOduration=4.618140195 podStartE2EDuration="4.618140195s" podCreationTimestamp="2025-09-30 08:16:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:16:05.586802031 +0000 UTC m=+789.039001298" watchObservedRunningTime="2025-09-30 08:16:05.618140195 +0000 UTC m=+789.070339462" Sep 30 08:16:07 crc kubenswrapper[4810]: I0930 08:16:07.295392 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fsz7d"] Sep 30 08:16:07 crc kubenswrapper[4810]: I0930 08:16:07.318096 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40f0c074-b5f2-464e-954d-44a022a65e85" path="/var/lib/kubelet/pods/40f0c074-b5f2-464e-954d-44a022a65e85/volumes" Sep 30 08:16:08 crc kubenswrapper[4810]: I0930 08:16:08.478144 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fsz7d" podUID="b73b1482-e35d-4784-8564-80f798b39727" containerName="registry-server" containerID="cri-o://fbe57a2f15bb0a2c9ca27ff4f8ae1af1557ceae18a755249c091536132797686" gracePeriod=2 Sep 30 08:16:09 crc kubenswrapper[4810]: I0930 08:16:09.485795 4810 generic.go:334] "Generic (PLEG): container finished" podID="b73b1482-e35d-4784-8564-80f798b39727" containerID="fbe57a2f15bb0a2c9ca27ff4f8ae1af1557ceae18a755249c091536132797686" exitCode=0 Sep 30 08:16:09 crc kubenswrapper[4810]: I0930 08:16:09.485844 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsz7d" event={"ID":"b73b1482-e35d-4784-8564-80f798b39727","Type":"ContainerDied","Data":"fbe57a2f15bb0a2c9ca27ff4f8ae1af1557ceae18a755249c091536132797686"} Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.519835 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-59tsf"] Sep 30 08:16:11 crc kubenswrapper[4810]: E0930 08:16:11.530809 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40f0c074-b5f2-464e-954d-44a022a65e85" containerName="extract-utilities" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.530867 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="40f0c074-b5f2-464e-954d-44a022a65e85" containerName="extract-utilities" Sep 30 08:16:11 crc kubenswrapper[4810]: E0930 08:16:11.530885 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40f0c074-b5f2-464e-954d-44a022a65e85" containerName="registry-server" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.530898 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="40f0c074-b5f2-464e-954d-44a022a65e85" containerName="registry-server" Sep 30 08:16:11 crc kubenswrapper[4810]: E0930 08:16:11.530919 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40f0c074-b5f2-464e-954d-44a022a65e85" containerName="extract-content" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.530932 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="40f0c074-b5f2-464e-954d-44a022a65e85" containerName="extract-content" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.531145 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="40f0c074-b5f2-464e-954d-44a022a65e85" containerName="registry-server" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.532991 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.548601 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-59tsf"] Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.619304 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdbbm\" (UniqueName: \"kubernetes.io/projected/42f207cb-9f55-417d-b619-82a58a1e1780-kube-api-access-zdbbm\") pod \"redhat-operators-59tsf\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.619362 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-utilities\") pod \"redhat-operators-59tsf\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.619456 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-catalog-content\") pod \"redhat-operators-59tsf\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.720432 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-catalog-content\") pod \"redhat-operators-59tsf\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.720553 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdbbm\" (UniqueName: \"kubernetes.io/projected/42f207cb-9f55-417d-b619-82a58a1e1780-kube-api-access-zdbbm\") pod \"redhat-operators-59tsf\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.720581 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-utilities\") pod \"redhat-operators-59tsf\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.721111 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-utilities\") pod \"redhat-operators-59tsf\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.721425 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-catalog-content\") pod \"redhat-operators-59tsf\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.752403 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdbbm\" (UniqueName: \"kubernetes.io/projected/42f207cb-9f55-417d-b619-82a58a1e1780-kube-api-access-zdbbm\") pod \"redhat-operators-59tsf\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:11 crc kubenswrapper[4810]: I0930 08:16:11.897364 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.064833 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.125617 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-utilities\") pod \"b73b1482-e35d-4784-8564-80f798b39727\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.125763 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-catalog-content\") pod \"b73b1482-e35d-4784-8564-80f798b39727\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.125795 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5c4g\" (UniqueName: \"kubernetes.io/projected/b73b1482-e35d-4784-8564-80f798b39727-kube-api-access-t5c4g\") pod \"b73b1482-e35d-4784-8564-80f798b39727\" (UID: \"b73b1482-e35d-4784-8564-80f798b39727\") " Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.126717 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-utilities" (OuterVolumeSpecName: "utilities") pod "b73b1482-e35d-4784-8564-80f798b39727" (UID: "b73b1482-e35d-4784-8564-80f798b39727"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.131398 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b73b1482-e35d-4784-8564-80f798b39727-kube-api-access-t5c4g" (OuterVolumeSpecName: "kube-api-access-t5c4g") pod "b73b1482-e35d-4784-8564-80f798b39727" (UID: "b73b1482-e35d-4784-8564-80f798b39727"). InnerVolumeSpecName "kube-api-access-t5c4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.184838 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b73b1482-e35d-4784-8564-80f798b39727" (UID: "b73b1482-e35d-4784-8564-80f798b39727"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.227250 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.227310 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b73b1482-e35d-4784-8564-80f798b39727-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.227326 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5c4g\" (UniqueName: \"kubernetes.io/projected/b73b1482-e35d-4784-8564-80f798b39727-kube-api-access-t5c4g\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.513788 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsz7d" event={"ID":"b73b1482-e35d-4784-8564-80f798b39727","Type":"ContainerDied","Data":"41e99b8218e4efbcfcecf9a0277b05aafe3e48036108b648ef908da59f630bd5"} Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.513885 4810 scope.go:117] "RemoveContainer" containerID="fbe57a2f15bb0a2c9ca27ff4f8ae1af1557ceae18a755249c091536132797686" Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.513907 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsz7d" Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.575428 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fsz7d"] Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.583612 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fsz7d"] Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.653645 4810 scope.go:117] "RemoveContainer" containerID="641dfed407850c9a2ded6c2eb3c7419260c5f8060e829ffe0f2541746ed9257c" Sep 30 08:16:12 crc kubenswrapper[4810]: I0930 08:16:12.712163 4810 scope.go:117] "RemoveContainer" containerID="7b181134fb76dbedcc0f3cb8069ccffdb3d85356337ea37cb688ed4fe2076f56" Sep 30 08:16:13 crc kubenswrapper[4810]: I0930 08:16:13.106348 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-59tsf"] Sep 30 08:16:13 crc kubenswrapper[4810]: I0930 08:16:13.315085 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b73b1482-e35d-4784-8564-80f798b39727" path="/var/lib/kubelet/pods/b73b1482-e35d-4784-8564-80f798b39727/volumes" Sep 30 08:16:13 crc kubenswrapper[4810]: I0930 08:16:13.525896 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" event={"ID":"ad0bc482-465c-433b-bd10-82ed2f79b6ad","Type":"ContainerStarted","Data":"181a03dbcfe53df75f0bc628ee3fc5d936eda7429daba898a5da49241488ec8f"} Sep 30 08:16:13 crc kubenswrapper[4810]: I0930 08:16:13.526140 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" Sep 30 08:16:13 crc kubenswrapper[4810]: I0930 08:16:13.528613 4810 generic.go:334] "Generic (PLEG): container finished" podID="4debe3c7-7058-4f37-863c-0c0e57818c86" containerID="b4e97c045abd55dd3be8459239584ea0dc88ea13a210c256a4a1fd894cf2247c" exitCode=0 Sep 30 08:16:13 crc kubenswrapper[4810]: I0930 08:16:13.528689 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wfqtl" event={"ID":"4debe3c7-7058-4f37-863c-0c0e57818c86","Type":"ContainerDied","Data":"b4e97c045abd55dd3be8459239584ea0dc88ea13a210c256a4a1fd894cf2247c"} Sep 30 08:16:13 crc kubenswrapper[4810]: I0930 08:16:13.530887 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59tsf" event={"ID":"42f207cb-9f55-417d-b619-82a58a1e1780","Type":"ContainerStarted","Data":"db16c1297b440c6bc0deca0b928ef20410f91d9f53d8c39449254a70856e1186"} Sep 30 08:16:13 crc kubenswrapper[4810]: I0930 08:16:13.553435 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" podStartSLOduration=2.39211 podStartE2EDuration="12.553400784s" podCreationTimestamp="2025-09-30 08:16:01 +0000 UTC" firstStartedPulling="2025-09-30 08:16:02.492246069 +0000 UTC m=+785.944445336" lastFinishedPulling="2025-09-30 08:16:12.653536853 +0000 UTC m=+796.105736120" observedRunningTime="2025-09-30 08:16:13.552191408 +0000 UTC m=+797.004390685" watchObservedRunningTime="2025-09-30 08:16:13.553400784 +0000 UTC m=+797.005600091" Sep 30 08:16:14 crc kubenswrapper[4810]: I0930 08:16:14.547308 4810 generic.go:334] "Generic (PLEG): container finished" podID="4debe3c7-7058-4f37-863c-0c0e57818c86" containerID="a966906d70a2dddb4937a82074884657e92c94c8c589eeffb07eb5792f985901" exitCode=0 Sep 30 08:16:14 crc kubenswrapper[4810]: I0930 08:16:14.547402 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wfqtl" event={"ID":"4debe3c7-7058-4f37-863c-0c0e57818c86","Type":"ContainerDied","Data":"a966906d70a2dddb4937a82074884657e92c94c8c589eeffb07eb5792f985901"} Sep 30 08:16:14 crc kubenswrapper[4810]: I0930 08:16:14.550582 4810 generic.go:334] "Generic (PLEG): container finished" podID="42f207cb-9f55-417d-b619-82a58a1e1780" containerID="e9aacc76466aca41f9a43aaaa9d425d68af9753eac232e59f037a35518aa3cfd" exitCode=0 Sep 30 08:16:14 crc kubenswrapper[4810]: I0930 08:16:14.551561 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59tsf" event={"ID":"42f207cb-9f55-417d-b619-82a58a1e1780","Type":"ContainerDied","Data":"e9aacc76466aca41f9a43aaaa9d425d68af9753eac232e59f037a35518aa3cfd"} Sep 30 08:16:15 crc kubenswrapper[4810]: I0930 08:16:15.563075 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59tsf" event={"ID":"42f207cb-9f55-417d-b619-82a58a1e1780","Type":"ContainerStarted","Data":"f3bc414391072a22f9bd00f07de0de10a6d522387777ac32a38024edcfd37ebd"} Sep 30 08:16:15 crc kubenswrapper[4810]: I0930 08:16:15.567999 4810 generic.go:334] "Generic (PLEG): container finished" podID="4debe3c7-7058-4f37-863c-0c0e57818c86" containerID="9f6c7291b01f3706a512a032457c4448edc8be3d5374d711fa0f35618ef67234" exitCode=0 Sep 30 08:16:15 crc kubenswrapper[4810]: I0930 08:16:15.568083 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wfqtl" event={"ID":"4debe3c7-7058-4f37-863c-0c0e57818c86","Type":"ContainerDied","Data":"9f6c7291b01f3706a512a032457c4448edc8be3d5374d711fa0f35618ef67234"} Sep 30 08:16:15 crc kubenswrapper[4810]: I0930 08:16:15.912045 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:16:15 crc kubenswrapper[4810]: I0930 08:16:15.912132 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:16:16 crc kubenswrapper[4810]: I0930 08:16:16.579400 4810 generic.go:334] "Generic (PLEG): container finished" podID="42f207cb-9f55-417d-b619-82a58a1e1780" containerID="f3bc414391072a22f9bd00f07de0de10a6d522387777ac32a38024edcfd37ebd" exitCode=0 Sep 30 08:16:16 crc kubenswrapper[4810]: I0930 08:16:16.579465 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59tsf" event={"ID":"42f207cb-9f55-417d-b619-82a58a1e1780","Type":"ContainerDied","Data":"f3bc414391072a22f9bd00f07de0de10a6d522387777ac32a38024edcfd37ebd"} Sep 30 08:16:16 crc kubenswrapper[4810]: I0930 08:16:16.587749 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wfqtl" event={"ID":"4debe3c7-7058-4f37-863c-0c0e57818c86","Type":"ContainerStarted","Data":"03d6a1774bc8cea22f6d501cfe176efb1dcc96c103ec07e74a1f50daec8524e0"} Sep 30 08:16:16 crc kubenswrapper[4810]: I0930 08:16:16.587803 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wfqtl" event={"ID":"4debe3c7-7058-4f37-863c-0c0e57818c86","Type":"ContainerStarted","Data":"00561252d88a073229e9bbfefe0080e233cedb3169b9debb9c5a4cc69f903b6e"} Sep 30 08:16:16 crc kubenswrapper[4810]: I0930 08:16:16.587817 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wfqtl" event={"ID":"4debe3c7-7058-4f37-863c-0c0e57818c86","Type":"ContainerStarted","Data":"46d2efe58a4022f9cc579ae1c52d9dd37b845f60751f6b2980321a964cf9428b"} Sep 30 08:16:16 crc kubenswrapper[4810]: I0930 08:16:16.587831 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wfqtl" event={"ID":"4debe3c7-7058-4f37-863c-0c0e57818c86","Type":"ContainerStarted","Data":"cc1169944406a2668a09f47013a3f3f63786a258ec29c8cd5c008546a51ca1f3"} Sep 30 08:16:16 crc kubenswrapper[4810]: I0930 08:16:16.587845 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wfqtl" event={"ID":"4debe3c7-7058-4f37-863c-0c0e57818c86","Type":"ContainerStarted","Data":"c6bd66dbe37bb3a6da1aabd4aaeb06c47d696a9fe2979741ed9232ab9488b2eb"} Sep 30 08:16:17 crc kubenswrapper[4810]: I0930 08:16:17.601805 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wfqtl" event={"ID":"4debe3c7-7058-4f37-863c-0c0e57818c86","Type":"ContainerStarted","Data":"48687cd973483f44ad6a6c314dc82b0fd375a45ffde33bc66612e7c6957930f1"} Sep 30 08:16:17 crc kubenswrapper[4810]: I0930 08:16:17.602650 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:17 crc kubenswrapper[4810]: I0930 08:16:17.604824 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59tsf" event={"ID":"42f207cb-9f55-417d-b619-82a58a1e1780","Type":"ContainerStarted","Data":"fd775179fde0fea9b5a8218f84f1e22ccd6353b9130deb99b09b50995304ddfe"} Sep 30 08:16:17 crc kubenswrapper[4810]: I0930 08:16:17.636778 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-wfqtl" podStartSLOduration=6.375546502 podStartE2EDuration="16.636755931s" podCreationTimestamp="2025-09-30 08:16:01 +0000 UTC" firstStartedPulling="2025-09-30 08:16:02.414293891 +0000 UTC m=+785.866493158" lastFinishedPulling="2025-09-30 08:16:12.67550332 +0000 UTC m=+796.127702587" observedRunningTime="2025-09-30 08:16:17.627850188 +0000 UTC m=+801.080049465" watchObservedRunningTime="2025-09-30 08:16:17.636755931 +0000 UTC m=+801.088955188" Sep 30 08:16:17 crc kubenswrapper[4810]: I0930 08:16:17.662599 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-59tsf" podStartSLOduration=4.241155348 podStartE2EDuration="6.662578942s" podCreationTimestamp="2025-09-30 08:16:11 +0000 UTC" firstStartedPulling="2025-09-30 08:16:14.554885303 +0000 UTC m=+798.007084590" lastFinishedPulling="2025-09-30 08:16:16.976308907 +0000 UTC m=+800.428508184" observedRunningTime="2025-09-30 08:16:17.66014116 +0000 UTC m=+801.112340437" watchObservedRunningTime="2025-09-30 08:16:17.662578942 +0000 UTC m=+801.114778209" Sep 30 08:16:21 crc kubenswrapper[4810]: I0930 08:16:21.717495 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-qr7xw" Sep 30 08:16:21 crc kubenswrapper[4810]: I0930 08:16:21.897693 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:21 crc kubenswrapper[4810]: I0930 08:16:21.897835 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:22 crc kubenswrapper[4810]: I0930 08:16:22.188666 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:22 crc kubenswrapper[4810]: I0930 08:16:22.218760 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-2cfm4" Sep 30 08:16:22 crc kubenswrapper[4810]: I0930 08:16:22.242085 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:22 crc kubenswrapper[4810]: I0930 08:16:22.943892 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-59tsf" podUID="42f207cb-9f55-417d-b619-82a58a1e1780" containerName="registry-server" probeResult="failure" output=< Sep 30 08:16:22 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Sep 30 08:16:22 crc kubenswrapper[4810]: > Sep 30 08:16:23 crc kubenswrapper[4810]: I0930 08:16:23.203979 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-gjfhs" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.371914 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-s9m98"] Sep 30 08:16:26 crc kubenswrapper[4810]: E0930 08:16:26.372659 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73b1482-e35d-4784-8564-80f798b39727" containerName="registry-server" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.372682 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73b1482-e35d-4784-8564-80f798b39727" containerName="registry-server" Sep 30 08:16:26 crc kubenswrapper[4810]: E0930 08:16:26.372714 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73b1482-e35d-4784-8564-80f798b39727" containerName="extract-utilities" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.372725 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73b1482-e35d-4784-8564-80f798b39727" containerName="extract-utilities" Sep 30 08:16:26 crc kubenswrapper[4810]: E0930 08:16:26.372739 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73b1482-e35d-4784-8564-80f798b39727" containerName="extract-content" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.372750 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73b1482-e35d-4784-8564-80f798b39727" containerName="extract-content" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.372951 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b73b1482-e35d-4784-8564-80f798b39727" containerName="registry-server" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.373665 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s9m98" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.375917 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.377144 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.385534 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-s9m98"] Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.450077 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhkhs\" (UniqueName: \"kubernetes.io/projected/0b883027-700c-48cd-a635-c471480871e2-kube-api-access-lhkhs\") pod \"openstack-operator-index-s9m98\" (UID: \"0b883027-700c-48cd-a635-c471480871e2\") " pod="openstack-operators/openstack-operator-index-s9m98" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.550945 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhkhs\" (UniqueName: \"kubernetes.io/projected/0b883027-700c-48cd-a635-c471480871e2-kube-api-access-lhkhs\") pod \"openstack-operator-index-s9m98\" (UID: \"0b883027-700c-48cd-a635-c471480871e2\") " pod="openstack-operators/openstack-operator-index-s9m98" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.576304 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhkhs\" (UniqueName: \"kubernetes.io/projected/0b883027-700c-48cd-a635-c471480871e2-kube-api-access-lhkhs\") pod \"openstack-operator-index-s9m98\" (UID: \"0b883027-700c-48cd-a635-c471480871e2\") " pod="openstack-operators/openstack-operator-index-s9m98" Sep 30 08:16:26 crc kubenswrapper[4810]: I0930 08:16:26.711076 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s9m98" Sep 30 08:16:27 crc kubenswrapper[4810]: I0930 08:16:27.191516 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-s9m98"] Sep 30 08:16:27 crc kubenswrapper[4810]: W0930 08:16:27.198248 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b883027_700c_48cd_a635_c471480871e2.slice/crio-49c96b6dbf7e1e298eb974967732bcb96d82086c036b6a12e31865545479a14e WatchSource:0}: Error finding container 49c96b6dbf7e1e298eb974967732bcb96d82086c036b6a12e31865545479a14e: Status 404 returned error can't find the container with id 49c96b6dbf7e1e298eb974967732bcb96d82086c036b6a12e31865545479a14e Sep 30 08:16:27 crc kubenswrapper[4810]: I0930 08:16:27.679338 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s9m98" event={"ID":"0b883027-700c-48cd-a635-c471480871e2","Type":"ContainerStarted","Data":"49c96b6dbf7e1e298eb974967732bcb96d82086c036b6a12e31865545479a14e"} Sep 30 08:16:29 crc kubenswrapper[4810]: I0930 08:16:29.736517 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-s9m98"] Sep 30 08:16:30 crc kubenswrapper[4810]: I0930 08:16:30.355535 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fqg48"] Sep 30 08:16:30 crc kubenswrapper[4810]: I0930 08:16:30.357143 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fqg48" Sep 30 08:16:30 crc kubenswrapper[4810]: I0930 08:16:30.360048 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-hn9hh" Sep 30 08:16:30 crc kubenswrapper[4810]: I0930 08:16:30.373105 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fqg48"] Sep 30 08:16:30 crc kubenswrapper[4810]: I0930 08:16:30.417172 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgwpd\" (UniqueName: \"kubernetes.io/projected/75091a5f-7967-4317-90b6-e195f197ecdb-kube-api-access-lgwpd\") pod \"openstack-operator-index-fqg48\" (UID: \"75091a5f-7967-4317-90b6-e195f197ecdb\") " pod="openstack-operators/openstack-operator-index-fqg48" Sep 30 08:16:30 crc kubenswrapper[4810]: I0930 08:16:30.519077 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgwpd\" (UniqueName: \"kubernetes.io/projected/75091a5f-7967-4317-90b6-e195f197ecdb-kube-api-access-lgwpd\") pod \"openstack-operator-index-fqg48\" (UID: \"75091a5f-7967-4317-90b6-e195f197ecdb\") " pod="openstack-operators/openstack-operator-index-fqg48" Sep 30 08:16:30 crc kubenswrapper[4810]: I0930 08:16:30.560188 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgwpd\" (UniqueName: \"kubernetes.io/projected/75091a5f-7967-4317-90b6-e195f197ecdb-kube-api-access-lgwpd\") pod \"openstack-operator-index-fqg48\" (UID: \"75091a5f-7967-4317-90b6-e195f197ecdb\") " pod="openstack-operators/openstack-operator-index-fqg48" Sep 30 08:16:30 crc kubenswrapper[4810]: I0930 08:16:30.688860 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fqg48" Sep 30 08:16:31 crc kubenswrapper[4810]: I0930 08:16:31.174644 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fqg48"] Sep 30 08:16:31 crc kubenswrapper[4810]: I0930 08:16:31.715777 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fqg48" event={"ID":"75091a5f-7967-4317-90b6-e195f197ecdb","Type":"ContainerStarted","Data":"1b472eb83a0c190cecf437f0a0d5c4f695be211b25daf3e1c73c8c1cccb00b8b"} Sep 30 08:16:31 crc kubenswrapper[4810]: I0930 08:16:31.974439 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:32 crc kubenswrapper[4810]: I0930 08:16:32.037027 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:32 crc kubenswrapper[4810]: I0930 08:16:32.200884 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-wfqtl" Sep 30 08:16:35 crc kubenswrapper[4810]: I0930 08:16:35.541190 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-59tsf"] Sep 30 08:16:35 crc kubenswrapper[4810]: I0930 08:16:35.541545 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-59tsf" podUID="42f207cb-9f55-417d-b619-82a58a1e1780" containerName="registry-server" containerID="cri-o://fd775179fde0fea9b5a8218f84f1e22ccd6353b9130deb99b09b50995304ddfe" gracePeriod=2 Sep 30 08:16:35 crc kubenswrapper[4810]: I0930 08:16:35.773467 4810 generic.go:334] "Generic (PLEG): container finished" podID="42f207cb-9f55-417d-b619-82a58a1e1780" containerID="fd775179fde0fea9b5a8218f84f1e22ccd6353b9130deb99b09b50995304ddfe" exitCode=0 Sep 30 08:16:35 crc kubenswrapper[4810]: I0930 08:16:35.773536 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59tsf" event={"ID":"42f207cb-9f55-417d-b619-82a58a1e1780","Type":"ContainerDied","Data":"fd775179fde0fea9b5a8218f84f1e22ccd6353b9130deb99b09b50995304ddfe"} Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.186059 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.319022 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-catalog-content\") pod \"42f207cb-9f55-417d-b619-82a58a1e1780\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.319122 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-utilities\") pod \"42f207cb-9f55-417d-b619-82a58a1e1780\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.319209 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdbbm\" (UniqueName: \"kubernetes.io/projected/42f207cb-9f55-417d-b619-82a58a1e1780-kube-api-access-zdbbm\") pod \"42f207cb-9f55-417d-b619-82a58a1e1780\" (UID: \"42f207cb-9f55-417d-b619-82a58a1e1780\") " Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.321196 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-utilities" (OuterVolumeSpecName: "utilities") pod "42f207cb-9f55-417d-b619-82a58a1e1780" (UID: "42f207cb-9f55-417d-b619-82a58a1e1780"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.328774 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42f207cb-9f55-417d-b619-82a58a1e1780-kube-api-access-zdbbm" (OuterVolumeSpecName: "kube-api-access-zdbbm") pod "42f207cb-9f55-417d-b619-82a58a1e1780" (UID: "42f207cb-9f55-417d-b619-82a58a1e1780"). InnerVolumeSpecName "kube-api-access-zdbbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.413236 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42f207cb-9f55-417d-b619-82a58a1e1780" (UID: "42f207cb-9f55-417d-b619-82a58a1e1780"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.424728 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdbbm\" (UniqueName: \"kubernetes.io/projected/42f207cb-9f55-417d-b619-82a58a1e1780-kube-api-access-zdbbm\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.424760 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.424772 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42f207cb-9f55-417d-b619-82a58a1e1780-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.781134 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59tsf" event={"ID":"42f207cb-9f55-417d-b619-82a58a1e1780","Type":"ContainerDied","Data":"db16c1297b440c6bc0deca0b928ef20410f91d9f53d8c39449254a70856e1186"} Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.781199 4810 scope.go:117] "RemoveContainer" containerID="fd775179fde0fea9b5a8218f84f1e22ccd6353b9130deb99b09b50995304ddfe" Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.781378 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59tsf" Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.821438 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-59tsf"] Sep 30 08:16:36 crc kubenswrapper[4810]: I0930 08:16:36.823940 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-59tsf"] Sep 30 08:16:37 crc kubenswrapper[4810]: I0930 08:16:37.349799 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42f207cb-9f55-417d-b619-82a58a1e1780" path="/var/lib/kubelet/pods/42f207cb-9f55-417d-b619-82a58a1e1780/volumes" Sep 30 08:16:37 crc kubenswrapper[4810]: I0930 08:16:37.455885 4810 scope.go:117] "RemoveContainer" containerID="f3bc414391072a22f9bd00f07de0de10a6d522387777ac32a38024edcfd37ebd" Sep 30 08:16:37 crc kubenswrapper[4810]: I0930 08:16:37.485620 4810 scope.go:117] "RemoveContainer" containerID="e9aacc76466aca41f9a43aaaa9d425d68af9753eac232e59f037a35518aa3cfd" Sep 30 08:16:38 crc kubenswrapper[4810]: I0930 08:16:38.800231 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fqg48" event={"ID":"75091a5f-7967-4317-90b6-e195f197ecdb","Type":"ContainerStarted","Data":"96d1735881ae85835b61dd5e9dacd0d976b14c913a908f23ade0acb549d44be3"} Sep 30 08:16:38 crc kubenswrapper[4810]: I0930 08:16:38.802025 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s9m98" event={"ID":"0b883027-700c-48cd-a635-c471480871e2","Type":"ContainerStarted","Data":"927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a"} Sep 30 08:16:38 crc kubenswrapper[4810]: I0930 08:16:38.802170 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-s9m98" podUID="0b883027-700c-48cd-a635-c471480871e2" containerName="registry-server" containerID="cri-o://927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a" gracePeriod=2 Sep 30 08:16:38 crc kubenswrapper[4810]: I0930 08:16:38.824328 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fqg48" podStartSLOduration=2.01007844 podStartE2EDuration="8.824295988s" podCreationTimestamp="2025-09-30 08:16:30 +0000 UTC" firstStartedPulling="2025-09-30 08:16:31.187493461 +0000 UTC m=+814.639692728" lastFinishedPulling="2025-09-30 08:16:38.001710969 +0000 UTC m=+821.453910276" observedRunningTime="2025-09-30 08:16:38.81810951 +0000 UTC m=+822.270308817" watchObservedRunningTime="2025-09-30 08:16:38.824295988 +0000 UTC m=+822.276495285" Sep 30 08:16:38 crc kubenswrapper[4810]: I0930 08:16:38.841065 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-s9m98" podStartSLOduration=2.048913714 podStartE2EDuration="12.84103244s" podCreationTimestamp="2025-09-30 08:16:26 +0000 UTC" firstStartedPulling="2025-09-30 08:16:27.208592604 +0000 UTC m=+810.660791891" lastFinishedPulling="2025-09-30 08:16:38.00071135 +0000 UTC m=+821.452910617" observedRunningTime="2025-09-30 08:16:38.837019704 +0000 UTC m=+822.289219021" watchObservedRunningTime="2025-09-30 08:16:38.84103244 +0000 UTC m=+822.293231737" Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.385961 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s9m98" Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.478665 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhkhs\" (UniqueName: \"kubernetes.io/projected/0b883027-700c-48cd-a635-c471480871e2-kube-api-access-lhkhs\") pod \"0b883027-700c-48cd-a635-c471480871e2\" (UID: \"0b883027-700c-48cd-a635-c471480871e2\") " Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.485479 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b883027-700c-48cd-a635-c471480871e2-kube-api-access-lhkhs" (OuterVolumeSpecName: "kube-api-access-lhkhs") pod "0b883027-700c-48cd-a635-c471480871e2" (UID: "0b883027-700c-48cd-a635-c471480871e2"). InnerVolumeSpecName "kube-api-access-lhkhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.579873 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhkhs\" (UniqueName: \"kubernetes.io/projected/0b883027-700c-48cd-a635-c471480871e2-kube-api-access-lhkhs\") on node \"crc\" DevicePath \"\"" Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.812376 4810 generic.go:334] "Generic (PLEG): container finished" podID="0b883027-700c-48cd-a635-c471480871e2" containerID="927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a" exitCode=0 Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.813051 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s9m98" Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.814507 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s9m98" event={"ID":"0b883027-700c-48cd-a635-c471480871e2","Type":"ContainerDied","Data":"927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a"} Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.814601 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s9m98" event={"ID":"0b883027-700c-48cd-a635-c471480871e2","Type":"ContainerDied","Data":"49c96b6dbf7e1e298eb974967732bcb96d82086c036b6a12e31865545479a14e"} Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.814646 4810 scope.go:117] "RemoveContainer" containerID="927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a" Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.841308 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-s9m98"] Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.847664 4810 scope.go:117] "RemoveContainer" containerID="927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a" Sep 30 08:16:39 crc kubenswrapper[4810]: E0930 08:16:39.848138 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a\": container with ID starting with 927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a not found: ID does not exist" containerID="927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a" Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.848180 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a"} err="failed to get container status \"927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a\": rpc error: code = NotFound desc = could not find container \"927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a\": container with ID starting with 927fecb4035e14c812872ab11b7176ee80677582eaaaf1693402710fd976376a not found: ID does not exist" Sep 30 08:16:39 crc kubenswrapper[4810]: I0930 08:16:39.854731 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-s9m98"] Sep 30 08:16:40 crc kubenswrapper[4810]: I0930 08:16:40.689829 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-fqg48" Sep 30 08:16:40 crc kubenswrapper[4810]: I0930 08:16:40.689939 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-fqg48" Sep 30 08:16:40 crc kubenswrapper[4810]: I0930 08:16:40.731203 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-fqg48" Sep 30 08:16:41 crc kubenswrapper[4810]: I0930 08:16:41.318396 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b883027-700c-48cd-a635-c471480871e2" path="/var/lib/kubelet/pods/0b883027-700c-48cd-a635-c471480871e2/volumes" Sep 30 08:16:45 crc kubenswrapper[4810]: I0930 08:16:45.911585 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:16:45 crc kubenswrapper[4810]: I0930 08:16:45.911962 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:16:50 crc kubenswrapper[4810]: I0930 08:16:50.733320 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-fqg48" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.670485 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv"] Sep 30 08:17:00 crc kubenswrapper[4810]: E0930 08:17:00.673599 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f207cb-9f55-417d-b619-82a58a1e1780" containerName="extract-utilities" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.673808 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f207cb-9f55-417d-b619-82a58a1e1780" containerName="extract-utilities" Sep 30 08:17:00 crc kubenswrapper[4810]: E0930 08:17:00.673952 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b883027-700c-48cd-a635-c471480871e2" containerName="registry-server" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.674088 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b883027-700c-48cd-a635-c471480871e2" containerName="registry-server" Sep 30 08:17:00 crc kubenswrapper[4810]: E0930 08:17:00.674223 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f207cb-9f55-417d-b619-82a58a1e1780" containerName="extract-content" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.674473 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f207cb-9f55-417d-b619-82a58a1e1780" containerName="extract-content" Sep 30 08:17:00 crc kubenswrapper[4810]: E0930 08:17:00.674643 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f207cb-9f55-417d-b619-82a58a1e1780" containerName="registry-server" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.674806 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f207cb-9f55-417d-b619-82a58a1e1780" containerName="registry-server" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.675226 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="42f207cb-9f55-417d-b619-82a58a1e1780" containerName="registry-server" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.675489 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b883027-700c-48cd-a635-c471480871e2" containerName="registry-server" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.677759 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.681561 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-gx5c4" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.689241 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv"] Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.701070 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-bundle\") pod \"0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.701166 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpt2k\" (UniqueName: \"kubernetes.io/projected/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-kube-api-access-kpt2k\") pod \"0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.701234 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-util\") pod \"0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.802955 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-util\") pod \"0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.803062 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-bundle\") pod \"0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.803127 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpt2k\" (UniqueName: \"kubernetes.io/projected/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-kube-api-access-kpt2k\") pod \"0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.803778 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-util\") pod \"0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.803980 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-bundle\") pod \"0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:00 crc kubenswrapper[4810]: I0930 08:17:00.837789 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpt2k\" (UniqueName: \"kubernetes.io/projected/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-kube-api-access-kpt2k\") pod \"0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:01 crc kubenswrapper[4810]: I0930 08:17:01.020059 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:01 crc kubenswrapper[4810]: I0930 08:17:01.252525 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv"] Sep 30 08:17:01 crc kubenswrapper[4810]: W0930 08:17:01.260415 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f1876f1_ed80_4dce_bdce_6a0c07a4f675.slice/crio-edcd36efa07ef5108824d6384e5d25ac27e5040dcdcdeb177c86343260cd70c4 WatchSource:0}: Error finding container edcd36efa07ef5108824d6384e5d25ac27e5040dcdcdeb177c86343260cd70c4: Status 404 returned error can't find the container with id edcd36efa07ef5108824d6384e5d25ac27e5040dcdcdeb177c86343260cd70c4 Sep 30 08:17:01 crc kubenswrapper[4810]: I0930 08:17:01.996971 4810 generic.go:334] "Generic (PLEG): container finished" podID="5f1876f1-ed80-4dce-bdce-6a0c07a4f675" containerID="809834247be885eda926469e3a02c65d528d3b1f89f7e3fc484d58af92d2c933" exitCode=0 Sep 30 08:17:01 crc kubenswrapper[4810]: I0930 08:17:01.997011 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" event={"ID":"5f1876f1-ed80-4dce-bdce-6a0c07a4f675","Type":"ContainerDied","Data":"809834247be885eda926469e3a02c65d528d3b1f89f7e3fc484d58af92d2c933"} Sep 30 08:17:01 crc kubenswrapper[4810]: I0930 08:17:01.997366 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" event={"ID":"5f1876f1-ed80-4dce-bdce-6a0c07a4f675","Type":"ContainerStarted","Data":"edcd36efa07ef5108824d6384e5d25ac27e5040dcdcdeb177c86343260cd70c4"} Sep 30 08:17:01 crc kubenswrapper[4810]: I0930 08:17:01.999141 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 08:17:03 crc kubenswrapper[4810]: I0930 08:17:03.007697 4810 generic.go:334] "Generic (PLEG): container finished" podID="5f1876f1-ed80-4dce-bdce-6a0c07a4f675" containerID="e49ffe580941a5fa72f4e56d41082675d7e24207c5f90e501b367d1fc7ef8e85" exitCode=0 Sep 30 08:17:03 crc kubenswrapper[4810]: I0930 08:17:03.008046 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" event={"ID":"5f1876f1-ed80-4dce-bdce-6a0c07a4f675","Type":"ContainerDied","Data":"e49ffe580941a5fa72f4e56d41082675d7e24207c5f90e501b367d1fc7ef8e85"} Sep 30 08:17:04 crc kubenswrapper[4810]: I0930 08:17:04.020962 4810 generic.go:334] "Generic (PLEG): container finished" podID="5f1876f1-ed80-4dce-bdce-6a0c07a4f675" containerID="e1e5472b6380bd13b08e60b2c65e07a7a9721620b27083180148f9ce69eccec4" exitCode=0 Sep 30 08:17:04 crc kubenswrapper[4810]: I0930 08:17:04.021019 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" event={"ID":"5f1876f1-ed80-4dce-bdce-6a0c07a4f675","Type":"ContainerDied","Data":"e1e5472b6380bd13b08e60b2c65e07a7a9721620b27083180148f9ce69eccec4"} Sep 30 08:17:05 crc kubenswrapper[4810]: I0930 08:17:05.348036 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:05 crc kubenswrapper[4810]: I0930 08:17:05.492731 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-util\") pod \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " Sep 30 08:17:05 crc kubenswrapper[4810]: I0930 08:17:05.493120 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpt2k\" (UniqueName: \"kubernetes.io/projected/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-kube-api-access-kpt2k\") pod \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " Sep 30 08:17:05 crc kubenswrapper[4810]: I0930 08:17:05.493372 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-bundle\") pod \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\" (UID: \"5f1876f1-ed80-4dce-bdce-6a0c07a4f675\") " Sep 30 08:17:05 crc kubenswrapper[4810]: I0930 08:17:05.494443 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-bundle" (OuterVolumeSpecName: "bundle") pod "5f1876f1-ed80-4dce-bdce-6a0c07a4f675" (UID: "5f1876f1-ed80-4dce-bdce-6a0c07a4f675"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:17:05 crc kubenswrapper[4810]: I0930 08:17:05.502932 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-kube-api-access-kpt2k" (OuterVolumeSpecName: "kube-api-access-kpt2k") pod "5f1876f1-ed80-4dce-bdce-6a0c07a4f675" (UID: "5f1876f1-ed80-4dce-bdce-6a0c07a4f675"). InnerVolumeSpecName "kube-api-access-kpt2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:17:05 crc kubenswrapper[4810]: I0930 08:17:05.523365 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-util" (OuterVolumeSpecName: "util") pod "5f1876f1-ed80-4dce-bdce-6a0c07a4f675" (UID: "5f1876f1-ed80-4dce-bdce-6a0c07a4f675"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:17:05 crc kubenswrapper[4810]: I0930 08:17:05.595792 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-util\") on node \"crc\" DevicePath \"\"" Sep 30 08:17:05 crc kubenswrapper[4810]: I0930 08:17:05.595851 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpt2k\" (UniqueName: \"kubernetes.io/projected/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-kube-api-access-kpt2k\") on node \"crc\" DevicePath \"\"" Sep 30 08:17:05 crc kubenswrapper[4810]: I0930 08:17:05.595874 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f1876f1-ed80-4dce-bdce-6a0c07a4f675-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:17:06 crc kubenswrapper[4810]: I0930 08:17:06.043509 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" event={"ID":"5f1876f1-ed80-4dce-bdce-6a0c07a4f675","Type":"ContainerDied","Data":"edcd36efa07ef5108824d6384e5d25ac27e5040dcdcdeb177c86343260cd70c4"} Sep 30 08:17:06 crc kubenswrapper[4810]: I0930 08:17:06.043867 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edcd36efa07ef5108824d6384e5d25ac27e5040dcdcdeb177c86343260cd70c4" Sep 30 08:17:06 crc kubenswrapper[4810]: I0930 08:17:06.043682 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.194245 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr"] Sep 30 08:17:13 crc kubenswrapper[4810]: E0930 08:17:13.194883 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f1876f1-ed80-4dce-bdce-6a0c07a4f675" containerName="util" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.194895 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f1876f1-ed80-4dce-bdce-6a0c07a4f675" containerName="util" Sep 30 08:17:13 crc kubenswrapper[4810]: E0930 08:17:13.194906 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f1876f1-ed80-4dce-bdce-6a0c07a4f675" containerName="pull" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.194911 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f1876f1-ed80-4dce-bdce-6a0c07a4f675" containerName="pull" Sep 30 08:17:13 crc kubenswrapper[4810]: E0930 08:17:13.194927 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f1876f1-ed80-4dce-bdce-6a0c07a4f675" containerName="extract" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.194933 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f1876f1-ed80-4dce-bdce-6a0c07a4f675" containerName="extract" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.195053 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f1876f1-ed80-4dce-bdce-6a0c07a4f675" containerName="extract" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.195660 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.198960 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-zzllh" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.230811 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr"] Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.327519 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7f7w\" (UniqueName: \"kubernetes.io/projected/154d7384-4a63-4a16-8400-903c4fe9eb8b-kube-api-access-c7f7w\") pod \"openstack-operator-controller-operator-6d66d9d974-d8fdr\" (UID: \"154d7384-4a63-4a16-8400-903c4fe9eb8b\") " pod="openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.429178 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7f7w\" (UniqueName: \"kubernetes.io/projected/154d7384-4a63-4a16-8400-903c4fe9eb8b-kube-api-access-c7f7w\") pod \"openstack-operator-controller-operator-6d66d9d974-d8fdr\" (UID: \"154d7384-4a63-4a16-8400-903c4fe9eb8b\") " pod="openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.450884 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7f7w\" (UniqueName: \"kubernetes.io/projected/154d7384-4a63-4a16-8400-903c4fe9eb8b-kube-api-access-c7f7w\") pod \"openstack-operator-controller-operator-6d66d9d974-d8fdr\" (UID: \"154d7384-4a63-4a16-8400-903c4fe9eb8b\") " pod="openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.511424 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr" Sep 30 08:17:13 crc kubenswrapper[4810]: I0930 08:17:13.797628 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr"] Sep 30 08:17:14 crc kubenswrapper[4810]: I0930 08:17:14.124533 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr" event={"ID":"154d7384-4a63-4a16-8400-903c4fe9eb8b","Type":"ContainerStarted","Data":"1aa567dd48ec59f23242778b21565f86cb439efcaf89d9fc2b036ad3862c7fea"} Sep 30 08:17:15 crc kubenswrapper[4810]: I0930 08:17:15.911169 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:17:15 crc kubenswrapper[4810]: I0930 08:17:15.911464 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:17:15 crc kubenswrapper[4810]: I0930 08:17:15.911520 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:17:15 crc kubenswrapper[4810]: I0930 08:17:15.912323 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2b0b0f4ca79c884d37aac25744a2d16466a6c0fbe29d05bbf5e9792b146f420"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:17:15 crc kubenswrapper[4810]: I0930 08:17:15.912386 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://b2b0b0f4ca79c884d37aac25744a2d16466a6c0fbe29d05bbf5e9792b146f420" gracePeriod=600 Sep 30 08:17:16 crc kubenswrapper[4810]: I0930 08:17:16.149129 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="b2b0b0f4ca79c884d37aac25744a2d16466a6c0fbe29d05bbf5e9792b146f420" exitCode=0 Sep 30 08:17:16 crc kubenswrapper[4810]: I0930 08:17:16.149178 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"b2b0b0f4ca79c884d37aac25744a2d16466a6c0fbe29d05bbf5e9792b146f420"} Sep 30 08:17:16 crc kubenswrapper[4810]: I0930 08:17:16.149216 4810 scope.go:117] "RemoveContainer" containerID="3e8713aca2fba3818ce4206358c20335307a8230c5f24759e0b223332c10c94d" Sep 30 08:17:18 crc kubenswrapper[4810]: I0930 08:17:18.163757 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr" event={"ID":"154d7384-4a63-4a16-8400-903c4fe9eb8b","Type":"ContainerStarted","Data":"691e92e72822acb6e98803c90288fdccbc16aa3a0ca1d44ec6ace883abeb5795"} Sep 30 08:17:18 crc kubenswrapper[4810]: I0930 08:17:18.165711 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"d1ca8f9a2402180e7c60714ae09e65a733987d89bb171e82c451bfe63cec7eed"} Sep 30 08:17:21 crc kubenswrapper[4810]: I0930 08:17:21.190814 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr" event={"ID":"154d7384-4a63-4a16-8400-903c4fe9eb8b","Type":"ContainerStarted","Data":"00f1ad6dd58535f14affc6bfe82765dc2ebe44e5461a7fc181535ca3a4c0123c"} Sep 30 08:17:21 crc kubenswrapper[4810]: I0930 08:17:21.191455 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr" Sep 30 08:17:21 crc kubenswrapper[4810]: I0930 08:17:21.239621 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr" podStartSLOduration=1.877714106 podStartE2EDuration="8.239601029s" podCreationTimestamp="2025-09-30 08:17:13 +0000 UTC" firstStartedPulling="2025-09-30 08:17:13.819554246 +0000 UTC m=+857.271753513" lastFinishedPulling="2025-09-30 08:17:20.181441169 +0000 UTC m=+863.633640436" observedRunningTime="2025-09-30 08:17:21.236540959 +0000 UTC m=+864.688740226" watchObservedRunningTime="2025-09-30 08:17:21.239601029 +0000 UTC m=+864.691800296" Sep 30 08:17:23 crc kubenswrapper[4810]: I0930 08:17:23.513992 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6d66d9d974-d8fdr" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.625042 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.626924 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.630935 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.632011 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.632404 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-fzsgm" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.634914 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-dgwtr" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.637471 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.638992 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.641868 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-67krc" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.652437 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.668323 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.679959 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.682243 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.692839 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-cbqht" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.763165 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.779701 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.779752 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.781565 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.785322 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.785795 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-l8klj" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.799039 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz9nz\" (UniqueName: \"kubernetes.io/projected/982a5d6f-5897-40d0-88a4-40a9f1ae9a5f-kube-api-access-tz9nz\") pod \"cinder-operator-controller-manager-644bddb6d8-6dhj8\" (UID: \"982a5d6f-5897-40d0-88a4-40a9f1ae9a5f\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.799123 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twntf\" (UniqueName: \"kubernetes.io/projected/da3d67ab-80be-486a-bc4d-d36ae6eb4756-kube-api-access-twntf\") pod \"glance-operator-controller-manager-84958c4d49-9wfgs\" (UID: \"da3d67ab-80be-486a-bc4d-d36ae6eb4756\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.799162 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbbmp\" (UniqueName: \"kubernetes.io/projected/6ad79408-8dc2-4e7e-aee3-54e3778a344e-kube-api-access-bbbmp\") pod \"barbican-operator-controller-manager-6ff8b75857-zzj8m\" (UID: \"6ad79408-8dc2-4e7e-aee3-54e3778a344e\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.799207 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb4wk\" (UniqueName: \"kubernetes.io/projected/02d77694-4bb0-4735-94c1-a3314ee634b9-kube-api-access-qb4wk\") pod \"designate-operator-controller-manager-84f4f7b77b-zfcrn\" (UID: \"02d77694-4bb0-4735-94c1-a3314ee634b9\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.801192 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.802736 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.804326 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.806031 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-kn86v" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.806217 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.810751 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.810762 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.812661 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.817853 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-xz475" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.822148 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-r4nzq" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.838481 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.842400 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.846926 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.848482 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.865551 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-zdbnn" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.866757 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.876082 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.877120 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.881586 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-sdgsj" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.893463 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.901049 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbbmp\" (UniqueName: \"kubernetes.io/projected/6ad79408-8dc2-4e7e-aee3-54e3778a344e-kube-api-access-bbbmp\") pod \"barbican-operator-controller-manager-6ff8b75857-zzj8m\" (UID: \"6ad79408-8dc2-4e7e-aee3-54e3778a344e\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.901126 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nlcc\" (UniqueName: \"kubernetes.io/projected/e801f35d-5d9f-48df-9a66-dd11ca14d26c-kube-api-access-7nlcc\") pod \"heat-operator-controller-manager-5d889d78cf-npvdf\" (UID: \"e801f35d-5d9f-48df-9a66-dd11ca14d26c\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.901176 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhgmb\" (UniqueName: \"kubernetes.io/projected/09ab9971-634a-40dd-88df-c1093fb66766-kube-api-access-zhgmb\") pod \"horizon-operator-controller-manager-9f4696d94-g76mw\" (UID: \"09ab9971-634a-40dd-88df-c1093fb66766\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.901213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb4wk\" (UniqueName: \"kubernetes.io/projected/02d77694-4bb0-4735-94c1-a3314ee634b9-kube-api-access-qb4wk\") pod \"designate-operator-controller-manager-84f4f7b77b-zfcrn\" (UID: \"02d77694-4bb0-4735-94c1-a3314ee634b9\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.901327 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz9nz\" (UniqueName: \"kubernetes.io/projected/982a5d6f-5897-40d0-88a4-40a9f1ae9a5f-kube-api-access-tz9nz\") pod \"cinder-operator-controller-manager-644bddb6d8-6dhj8\" (UID: \"982a5d6f-5897-40d0-88a4-40a9f1ae9a5f\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.901387 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twntf\" (UniqueName: \"kubernetes.io/projected/da3d67ab-80be-486a-bc4d-d36ae6eb4756-kube-api-access-twntf\") pod \"glance-operator-controller-manager-84958c4d49-9wfgs\" (UID: \"da3d67ab-80be-486a-bc4d-d36ae6eb4756\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.918411 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.925535 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz9nz\" (UniqueName: \"kubernetes.io/projected/982a5d6f-5897-40d0-88a4-40a9f1ae9a5f-kube-api-access-tz9nz\") pod \"cinder-operator-controller-manager-644bddb6d8-6dhj8\" (UID: \"982a5d6f-5897-40d0-88a4-40a9f1ae9a5f\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.925773 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbbmp\" (UniqueName: \"kubernetes.io/projected/6ad79408-8dc2-4e7e-aee3-54e3778a344e-kube-api-access-bbbmp\") pod \"barbican-operator-controller-manager-6ff8b75857-zzj8m\" (UID: \"6ad79408-8dc2-4e7e-aee3-54e3778a344e\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.934453 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twntf\" (UniqueName: \"kubernetes.io/projected/da3d67ab-80be-486a-bc4d-d36ae6eb4756-kube-api-access-twntf\") pod \"glance-operator-controller-manager-84958c4d49-9wfgs\" (UID: \"da3d67ab-80be-486a-bc4d-d36ae6eb4756\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.940708 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb4wk\" (UniqueName: \"kubernetes.io/projected/02d77694-4bb0-4735-94c1-a3314ee634b9-kube-api-access-qb4wk\") pod \"designate-operator-controller-manager-84f4f7b77b-zfcrn\" (UID: \"02d77694-4bb0-4735-94c1-a3314ee634b9\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.946065 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-wnngc"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.947491 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-wnngc" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.948876 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-wnngc"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.953026 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-fx2r4" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.956790 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.958584 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.963529 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.983038 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt"] Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.983404 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-hhnp2" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.984181 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.994127 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-ngkrm" Sep 30 08:17:45 crc kubenswrapper[4810]: I0930 08:17:45.997462 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.002116 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmljr\" (UniqueName: \"kubernetes.io/projected/8b3a8007-4969-4c86-a03a-ce633b3c31aa-kube-api-access-pmljr\") pod \"infra-operator-controller-manager-7d857cc749-v9pjf\" (UID: \"8b3a8007-4969-4c86-a03a-ce633b3c31aa\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.002178 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b3a8007-4969-4c86-a03a-ce633b3c31aa-cert\") pod \"infra-operator-controller-manager-7d857cc749-v9pjf\" (UID: \"8b3a8007-4969-4c86-a03a-ce633b3c31aa\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.002221 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nlcc\" (UniqueName: \"kubernetes.io/projected/e801f35d-5d9f-48df-9a66-dd11ca14d26c-kube-api-access-7nlcc\") pod \"heat-operator-controller-manager-5d889d78cf-npvdf\" (UID: \"e801f35d-5d9f-48df-9a66-dd11ca14d26c\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.002260 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhgmb\" (UniqueName: \"kubernetes.io/projected/09ab9971-634a-40dd-88df-c1093fb66766-kube-api-access-zhgmb\") pod \"horizon-operator-controller-manager-9f4696d94-g76mw\" (UID: \"09ab9971-634a-40dd-88df-c1093fb66766\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.002301 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brlf6\" (UniqueName: \"kubernetes.io/projected/52a2d49f-0736-4cac-8363-1a71fa6c1dd4-kube-api-access-brlf6\") pod \"keystone-operator-controller-manager-5bd55b4bff-vz9pd\" (UID: \"52a2d49f-0736-4cac-8363-1a71fa6c1dd4\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.002345 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jmkb\" (UniqueName: \"kubernetes.io/projected/8c5f7678-0f63-4323-910c-227040fdac66-kube-api-access-8jmkb\") pod \"ironic-operator-controller-manager-7975b88857-hd4x4\" (UID: \"8c5f7678-0f63-4323-910c-227040fdac66\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.002380 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c67ql\" (UniqueName: \"kubernetes.io/projected/df615184-edc5-4d93-b0ae-b414d9eb0ce1-kube-api-access-c67ql\") pod \"manila-operator-controller-manager-6d68dbc695-8bwwd\" (UID: \"df615184-edc5-4d93-b0ae-b414d9eb0ce1\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.015796 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.028636 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.036358 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nlcc\" (UniqueName: \"kubernetes.io/projected/e801f35d-5d9f-48df-9a66-dd11ca14d26c-kube-api-access-7nlcc\") pod \"heat-operator-controller-manager-5d889d78cf-npvdf\" (UID: \"e801f35d-5d9f-48df-9a66-dd11ca14d26c\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.041542 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.048937 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.049093 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.052884 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-kvsr6" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.061179 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhgmb\" (UniqueName: \"kubernetes.io/projected/09ab9971-634a-40dd-88df-c1093fb66766-kube-api-access-zhgmb\") pod \"horizon-operator-controller-manager-9f4696d94-g76mw\" (UID: \"09ab9971-634a-40dd-88df-c1093fb66766\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.073523 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.079517 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.080933 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.094829 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-v4hth" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.095055 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.095386 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.104745 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmljr\" (UniqueName: \"kubernetes.io/projected/8b3a8007-4969-4c86-a03a-ce633b3c31aa-kube-api-access-pmljr\") pod \"infra-operator-controller-manager-7d857cc749-v9pjf\" (UID: \"8b3a8007-4969-4c86-a03a-ce633b3c31aa\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.104787 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b3a8007-4969-4c86-a03a-ce633b3c31aa-cert\") pod \"infra-operator-controller-manager-7d857cc749-v9pjf\" (UID: \"8b3a8007-4969-4c86-a03a-ce633b3c31aa\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.104832 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd6cf\" (UniqueName: \"kubernetes.io/projected/b32751fb-95c0-4295-937a-ce7f82562863-kube-api-access-cd6cf\") pod \"nova-operator-controller-manager-c7c776c96-v58wt\" (UID: \"b32751fb-95c0-4295-937a-ce7f82562863\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.104855 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brlf6\" (UniqueName: \"kubernetes.io/projected/52a2d49f-0736-4cac-8363-1a71fa6c1dd4-kube-api-access-brlf6\") pod \"keystone-operator-controller-manager-5bd55b4bff-vz9pd\" (UID: \"52a2d49f-0736-4cac-8363-1a71fa6c1dd4\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.104879 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv8nf\" (UniqueName: \"kubernetes.io/projected/ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e-kube-api-access-sv8nf\") pod \"neutron-operator-controller-manager-64d7b59854-t6wwm\" (UID: \"ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.104903 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jmkb\" (UniqueName: \"kubernetes.io/projected/8c5f7678-0f63-4323-910c-227040fdac66-kube-api-access-8jmkb\") pod \"ironic-operator-controller-manager-7975b88857-hd4x4\" (UID: \"8c5f7678-0f63-4323-910c-227040fdac66\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.104928 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c67ql\" (UniqueName: \"kubernetes.io/projected/df615184-edc5-4d93-b0ae-b414d9eb0ce1-kube-api-access-c67ql\") pod \"manila-operator-controller-manager-6d68dbc695-8bwwd\" (UID: \"df615184-edc5-4d93-b0ae-b414d9eb0ce1\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.104947 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jbhv\" (UniqueName: \"kubernetes.io/projected/8a8a969d-2c5d-4229-b01d-cabcf3aec35c-kube-api-access-2jbhv\") pod \"mariadb-operator-controller-manager-88c7-wnngc\" (UID: \"8a8a969d-2c5d-4229-b01d-cabcf3aec35c\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-wnngc" Sep 30 08:17:46 crc kubenswrapper[4810]: E0930 08:17:46.105233 4810 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Sep 30 08:17:46 crc kubenswrapper[4810]: E0930 08:17:46.105291 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b3a8007-4969-4c86-a03a-ce633b3c31aa-cert podName:8b3a8007-4969-4c86-a03a-ce633b3c31aa nodeName:}" failed. No retries permitted until 2025-09-30 08:17:46.605276554 +0000 UTC m=+890.057475821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8b3a8007-4969-4c86-a03a-ce633b3c31aa-cert") pod "infra-operator-controller-manager-7d857cc749-v9pjf" (UID: "8b3a8007-4969-4c86-a03a-ce633b3c31aa") : secret "infra-operator-webhook-server-cert" not found Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.107611 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.114456 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.115483 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.119599 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-8m74x" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.125856 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.127048 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.127323 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.134747 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-pq7qd" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.142345 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmljr\" (UniqueName: \"kubernetes.io/projected/8b3a8007-4969-4c86-a03a-ce633b3c31aa-kube-api-access-pmljr\") pod \"infra-operator-controller-manager-7d857cc749-v9pjf\" (UID: \"8b3a8007-4969-4c86-a03a-ce633b3c31aa\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.153319 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jmkb\" (UniqueName: \"kubernetes.io/projected/8c5f7678-0f63-4323-910c-227040fdac66-kube-api-access-8jmkb\") pod \"ironic-operator-controller-manager-7975b88857-hd4x4\" (UID: \"8c5f7678-0f63-4323-910c-227040fdac66\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.156066 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brlf6\" (UniqueName: \"kubernetes.io/projected/52a2d49f-0736-4cac-8363-1a71fa6c1dd4-kube-api-access-brlf6\") pod \"keystone-operator-controller-manager-5bd55b4bff-vz9pd\" (UID: \"52a2d49f-0736-4cac-8363-1a71fa6c1dd4\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.164144 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.189673 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c67ql\" (UniqueName: \"kubernetes.io/projected/df615184-edc5-4d93-b0ae-b414d9eb0ce1-kube-api-access-c67ql\") pod \"manila-operator-controller-manager-6d68dbc695-8bwwd\" (UID: \"df615184-edc5-4d93-b0ae-b414d9eb0ce1\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.197732 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.206203 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv8nf\" (UniqueName: \"kubernetes.io/projected/ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e-kube-api-access-sv8nf\") pod \"neutron-operator-controller-manager-64d7b59854-t6wwm\" (UID: \"ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.206283 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jbhv\" (UniqueName: \"kubernetes.io/projected/8a8a969d-2c5d-4229-b01d-cabcf3aec35c-kube-api-access-2jbhv\") pod \"mariadb-operator-controller-manager-88c7-wnngc\" (UID: \"8a8a969d-2c5d-4229-b01d-cabcf3aec35c\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-wnngc" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.206345 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58s4z\" (UniqueName: \"kubernetes.io/projected/0f34a7d1-aa5e-4b55-a75d-5366e196e742-kube-api-access-58s4z\") pod \"octavia-operator-controller-manager-76fcc6dc7c-sndbm\" (UID: \"0f34a7d1-aa5e-4b55-a75d-5366e196e742\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.206403 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75l7b\" (UniqueName: \"kubernetes.io/projected/4e9c80fa-52fc-41db-a0be-babb0bd14b56-kube-api-access-75l7b\") pod \"openstack-baremetal-operator-controller-manager-6d776955-9gsjq\" (UID: \"4e9c80fa-52fc-41db-a0be-babb0bd14b56\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.206435 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4e9c80fa-52fc-41db-a0be-babb0bd14b56-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-9gsjq\" (UID: \"4e9c80fa-52fc-41db-a0be-babb0bd14b56\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.206463 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd6cf\" (UniqueName: \"kubernetes.io/projected/b32751fb-95c0-4295-937a-ce7f82562863-kube-api-access-cd6cf\") pod \"nova-operator-controller-manager-c7c776c96-v58wt\" (UID: \"b32751fb-95c0-4295-937a-ce7f82562863\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.209658 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.277278 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jbhv\" (UniqueName: \"kubernetes.io/projected/8a8a969d-2c5d-4229-b01d-cabcf3aec35c-kube-api-access-2jbhv\") pod \"mariadb-operator-controller-manager-88c7-wnngc\" (UID: \"8a8a969d-2c5d-4229-b01d-cabcf3aec35c\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-wnngc" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.279005 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv8nf\" (UniqueName: \"kubernetes.io/projected/ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e-kube-api-access-sv8nf\") pod \"neutron-operator-controller-manager-64d7b59854-t6wwm\" (UID: \"ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.279763 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.280320 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd6cf\" (UniqueName: \"kubernetes.io/projected/b32751fb-95c0-4295-937a-ce7f82562863-kube-api-access-cd6cf\") pod \"nova-operator-controller-manager-c7c776c96-v58wt\" (UID: \"b32751fb-95c0-4295-937a-ce7f82562863\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.290967 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.307347 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m48d4\" (UniqueName: \"kubernetes.io/projected/11f06c30-1cef-40f3-a493-8be235930b71-kube-api-access-m48d4\") pod \"placement-operator-controller-manager-589c58c6c-t24hl\" (UID: \"11f06c30-1cef-40f3-a493-8be235930b71\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.307395 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvh42\" (UniqueName: \"kubernetes.io/projected/24044399-d033-436a-ad32-1e1aefe76626-kube-api-access-nvh42\") pod \"ovn-operator-controller-manager-9976ff44c-s785j\" (UID: \"24044399-d033-436a-ad32-1e1aefe76626\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.307426 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58s4z\" (UniqueName: \"kubernetes.io/projected/0f34a7d1-aa5e-4b55-a75d-5366e196e742-kube-api-access-58s4z\") pod \"octavia-operator-controller-manager-76fcc6dc7c-sndbm\" (UID: \"0f34a7d1-aa5e-4b55-a75d-5366e196e742\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.307481 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75l7b\" (UniqueName: \"kubernetes.io/projected/4e9c80fa-52fc-41db-a0be-babb0bd14b56-kube-api-access-75l7b\") pod \"openstack-baremetal-operator-controller-manager-6d776955-9gsjq\" (UID: \"4e9c80fa-52fc-41db-a0be-babb0bd14b56\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.307509 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4e9c80fa-52fc-41db-a0be-babb0bd14b56-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-9gsjq\" (UID: \"4e9c80fa-52fc-41db-a0be-babb0bd14b56\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:17:46 crc kubenswrapper[4810]: E0930 08:17:46.307631 4810 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 08:17:46 crc kubenswrapper[4810]: E0930 08:17:46.307668 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4e9c80fa-52fc-41db-a0be-babb0bd14b56-cert podName:4e9c80fa-52fc-41db-a0be-babb0bd14b56 nodeName:}" failed. No retries permitted until 2025-09-30 08:17:46.807655463 +0000 UTC m=+890.259854730 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4e9c80fa-52fc-41db-a0be-babb0bd14b56-cert") pod "openstack-baremetal-operator-controller-manager-6d776955-9gsjq" (UID: "4e9c80fa-52fc-41db-a0be-babb0bd14b56") : secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.314544 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-wnngc" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.338754 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58s4z\" (UniqueName: \"kubernetes.io/projected/0f34a7d1-aa5e-4b55-a75d-5366e196e742-kube-api-access-58s4z\") pod \"octavia-operator-controller-manager-76fcc6dc7c-sndbm\" (UID: \"0f34a7d1-aa5e-4b55-a75d-5366e196e742\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.347118 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.348451 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.353348 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.355097 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.358739 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.359409 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-w777h" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.359714 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-zwbgx" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.360785 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75l7b\" (UniqueName: \"kubernetes.io/projected/4e9c80fa-52fc-41db-a0be-babb0bd14b56-kube-api-access-75l7b\") pod \"openstack-baremetal-operator-controller-manager-6d776955-9gsjq\" (UID: \"4e9c80fa-52fc-41db-a0be-babb0bd14b56\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.380870 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.403792 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.411907 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m48d4\" (UniqueName: \"kubernetes.io/projected/11f06c30-1cef-40f3-a493-8be235930b71-kube-api-access-m48d4\") pod \"placement-operator-controller-manager-589c58c6c-t24hl\" (UID: \"11f06c30-1cef-40f3-a493-8be235930b71\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.411949 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvh42\" (UniqueName: \"kubernetes.io/projected/24044399-d033-436a-ad32-1e1aefe76626-kube-api-access-nvh42\") pod \"ovn-operator-controller-manager-9976ff44c-s785j\" (UID: \"24044399-d033-436a-ad32-1e1aefe76626\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.435830 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvh42\" (UniqueName: \"kubernetes.io/projected/24044399-d033-436a-ad32-1e1aefe76626-kube-api-access-nvh42\") pod \"ovn-operator-controller-manager-9976ff44c-s785j\" (UID: \"24044399-d033-436a-ad32-1e1aefe76626\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.441411 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-g8s55"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.442704 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.445839 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-2bz9m" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.445955 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m48d4\" (UniqueName: \"kubernetes.io/projected/11f06c30-1cef-40f3-a493-8be235930b71-kube-api-access-m48d4\") pod \"placement-operator-controller-manager-589c58c6c-t24hl\" (UID: \"11f06c30-1cef-40f3-a493-8be235930b71\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.446304 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.451735 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.451814 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.471976 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-g8s55"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.514060 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4dt4\" (UniqueName: \"kubernetes.io/projected/0cc20b2d-4504-4f46-b0a0-5aee4b2201a3-kube-api-access-g4dt4\") pod \"telemetry-operator-controller-manager-b8d54b5d7-g8ls7\" (UID: \"0cc20b2d-4504-4f46-b0a0-5aee4b2201a3\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.514145 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmqgp\" (UniqueName: \"kubernetes.io/projected/3bbbc1cd-370d-468f-8985-cdf9219ca326-kube-api-access-qmqgp\") pod \"swift-operator-controller-manager-bc7dc7bd9-tc2wc\" (UID: \"3bbbc1cd-370d-468f-8985-cdf9219ca326\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.520672 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.521834 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.522784 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.524640 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.529076 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-5z8d8" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.546782 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.591609 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.592648 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.592733 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.600971 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-djmr9" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.602063 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.606433 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.607522 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.611865 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-nwc4p" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.616191 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4dt4\" (UniqueName: \"kubernetes.io/projected/0cc20b2d-4504-4f46-b0a0-5aee4b2201a3-kube-api-access-g4dt4\") pod \"telemetry-operator-controller-manager-b8d54b5d7-g8ls7\" (UID: \"0cc20b2d-4504-4f46-b0a0-5aee4b2201a3\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.616238 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b3a8007-4969-4c86-a03a-ce633b3c31aa-cert\") pod \"infra-operator-controller-manager-7d857cc749-v9pjf\" (UID: \"8b3a8007-4969-4c86-a03a-ce633b3c31aa\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.616295 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tbfw\" (UniqueName: \"kubernetes.io/projected/7fe8706c-6dfa-4a2f-876b-4cded24209bf-kube-api-access-6tbfw\") pod \"test-operator-controller-manager-f66b554c6-g8s55\" (UID: \"7fe8706c-6dfa-4a2f-876b-4cded24209bf\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.616327 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmqgp\" (UniqueName: \"kubernetes.io/projected/3bbbc1cd-370d-468f-8985-cdf9219ca326-kube-api-access-qmqgp\") pod \"swift-operator-controller-manager-bc7dc7bd9-tc2wc\" (UID: \"3bbbc1cd-370d-468f-8985-cdf9219ca326\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.622252 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.639512 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmqgp\" (UniqueName: \"kubernetes.io/projected/3bbbc1cd-370d-468f-8985-cdf9219ca326-kube-api-access-qmqgp\") pod \"swift-operator-controller-manager-bc7dc7bd9-tc2wc\" (UID: \"3bbbc1cd-370d-468f-8985-cdf9219ca326\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.640144 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4dt4\" (UniqueName: \"kubernetes.io/projected/0cc20b2d-4504-4f46-b0a0-5aee4b2201a3-kube-api-access-g4dt4\") pod \"telemetry-operator-controller-manager-b8d54b5d7-g8ls7\" (UID: \"0cc20b2d-4504-4f46-b0a0-5aee4b2201a3\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.645050 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b3a8007-4969-4c86-a03a-ce633b3c31aa-cert\") pod \"infra-operator-controller-manager-7d857cc749-v9pjf\" (UID: \"8b3a8007-4969-4c86-a03a-ce633b3c31aa\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.694678 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.717063 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tbfw\" (UniqueName: \"kubernetes.io/projected/7fe8706c-6dfa-4a2f-876b-4cded24209bf-kube-api-access-6tbfw\") pod \"test-operator-controller-manager-f66b554c6-g8s55\" (UID: \"7fe8706c-6dfa-4a2f-876b-4cded24209bf\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.717106 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5kmt\" (UniqueName: \"kubernetes.io/projected/b63222b9-1cc3-4a55-a498-ce538cea80b9-kube-api-access-n5kmt\") pod \"watcher-operator-controller-manager-59446d9d9-kb958\" (UID: \"b63222b9-1cc3-4a55-a498-ce538cea80b9\") " pod="openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.717145 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffjnq\" (UniqueName: \"kubernetes.io/projected/3df3a4ac-fc49-404f-aaba-cc8c053365c4-kube-api-access-ffjnq\") pod \"openstack-operator-controller-manager-64647d55fc-6z69x\" (UID: \"3df3a4ac-fc49-404f-aaba-cc8c053365c4\") " pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.717222 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vmjr\" (UniqueName: \"kubernetes.io/projected/bf4551e1-3981-407f-ba07-bff4db2b6d6d-kube-api-access-9vmjr\") pod \"rabbitmq-cluster-operator-manager-79d8469568-js8pw\" (UID: \"bf4551e1-3981-407f-ba07-bff4db2b6d6d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.717244 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3df3a4ac-fc49-404f-aaba-cc8c053365c4-cert\") pod \"openstack-operator-controller-manager-64647d55fc-6z69x\" (UID: \"3df3a4ac-fc49-404f-aaba-cc8c053365c4\") " pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.724550 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.741106 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.748328 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tbfw\" (UniqueName: \"kubernetes.io/projected/7fe8706c-6dfa-4a2f-876b-4cded24209bf-kube-api-access-6tbfw\") pod \"test-operator-controller-manager-f66b554c6-g8s55\" (UID: \"7fe8706c-6dfa-4a2f-876b-4cded24209bf\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.773920 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.779082 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.820883 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vmjr\" (UniqueName: \"kubernetes.io/projected/bf4551e1-3981-407f-ba07-bff4db2b6d6d-kube-api-access-9vmjr\") pod \"rabbitmq-cluster-operator-manager-79d8469568-js8pw\" (UID: \"bf4551e1-3981-407f-ba07-bff4db2b6d6d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.820921 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3df3a4ac-fc49-404f-aaba-cc8c053365c4-cert\") pod \"openstack-operator-controller-manager-64647d55fc-6z69x\" (UID: \"3df3a4ac-fc49-404f-aaba-cc8c053365c4\") " pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.820958 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4e9c80fa-52fc-41db-a0be-babb0bd14b56-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-9gsjq\" (UID: \"4e9c80fa-52fc-41db-a0be-babb0bd14b56\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.820979 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5kmt\" (UniqueName: \"kubernetes.io/projected/b63222b9-1cc3-4a55-a498-ce538cea80b9-kube-api-access-n5kmt\") pod \"watcher-operator-controller-manager-59446d9d9-kb958\" (UID: \"b63222b9-1cc3-4a55-a498-ce538cea80b9\") " pod="openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.821018 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffjnq\" (UniqueName: \"kubernetes.io/projected/3df3a4ac-fc49-404f-aaba-cc8c053365c4-kube-api-access-ffjnq\") pod \"openstack-operator-controller-manager-64647d55fc-6z69x\" (UID: \"3df3a4ac-fc49-404f-aaba-cc8c053365c4\") " pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" Sep 30 08:17:46 crc kubenswrapper[4810]: E0930 08:17:46.822292 4810 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 08:17:46 crc kubenswrapper[4810]: E0930 08:17:46.822373 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4e9c80fa-52fc-41db-a0be-babb0bd14b56-cert podName:4e9c80fa-52fc-41db-a0be-babb0bd14b56 nodeName:}" failed. No retries permitted until 2025-09-30 08:17:47.822351514 +0000 UTC m=+891.274550781 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4e9c80fa-52fc-41db-a0be-babb0bd14b56-cert") pod "openstack-baremetal-operator-controller-manager-6d776955-9gsjq" (UID: "4e9c80fa-52fc-41db-a0be-babb0bd14b56") : secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.827982 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3df3a4ac-fc49-404f-aaba-cc8c053365c4-cert\") pod \"openstack-operator-controller-manager-64647d55fc-6z69x\" (UID: \"3df3a4ac-fc49-404f-aaba-cc8c053365c4\") " pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.851049 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffjnq\" (UniqueName: \"kubernetes.io/projected/3df3a4ac-fc49-404f-aaba-cc8c053365c4-kube-api-access-ffjnq\") pod \"openstack-operator-controller-manager-64647d55fc-6z69x\" (UID: \"3df3a4ac-fc49-404f-aaba-cc8c053365c4\") " pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.856515 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5kmt\" (UniqueName: \"kubernetes.io/projected/b63222b9-1cc3-4a55-a498-ce538cea80b9-kube-api-access-n5kmt\") pod \"watcher-operator-controller-manager-59446d9d9-kb958\" (UID: \"b63222b9-1cc3-4a55-a498-ce538cea80b9\") " pod="openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.856812 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.870117 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vmjr\" (UniqueName: \"kubernetes.io/projected/bf4551e1-3981-407f-ba07-bff4db2b6d6d-kube-api-access-9vmjr\") pod \"rabbitmq-cluster-operator-manager-79d8469568-js8pw\" (UID: \"bf4551e1-3981-407f-ba07-bff4db2b6d6d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.947764 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8"] Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.951582 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" Sep 30 08:17:46 crc kubenswrapper[4810]: I0930 08:17:46.964963 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.382868 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.396060 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.406455 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs" event={"ID":"da3d67ab-80be-486a-bc4d-d36ae6eb4756","Type":"ContainerStarted","Data":"b09d4ca59f627d4beec845af48a2259b4478d6a358025ed0c70d59a1d59604c9"} Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.408648 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw" event={"ID":"09ab9971-634a-40dd-88df-c1093fb66766","Type":"ContainerStarted","Data":"718b57a6cce549d5444e39d1dd47bc59d9a82e2c7c8f7dff0d77beecb3349ddd"} Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.420540 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8" event={"ID":"982a5d6f-5897-40d0-88a4-40a9f1ae9a5f","Type":"ContainerStarted","Data":"d132d196e35d386f34a59109b941fa2395ea771f6efd924ccae6696e8c4244f7"} Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.423130 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m" event={"ID":"6ad79408-8dc2-4e7e-aee3-54e3778a344e","Type":"ContainerStarted","Data":"b02a2397fc7b0d5113076421969731358b9cae74345d74bf0804934a4033882b"} Sep 30 08:17:47 crc kubenswrapper[4810]: W0930 08:17:47.449136 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf615184_edc5_4d93_b0ae_b414d9eb0ce1.slice/crio-7f70cc06e40267162f23594ea4380ec174af82ea910d71c7bc625f418824cc16 WatchSource:0}: Error finding container 7f70cc06e40267162f23594ea4380ec174af82ea910d71c7bc625f418824cc16: Status 404 returned error can't find the container with id 7f70cc06e40267162f23594ea4380ec174af82ea910d71c7bc625f418824cc16 Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.459484 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.466591 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.471361 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.475786 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.648320 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j"] Sep 30 08:17:47 crc kubenswrapper[4810]: W0930 08:17:47.649389 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24044399_d033_436a_ad32_1e1aefe76626.slice/crio-151eab162fb07c7bec047f0b58f05784e69cfb48ede30f78e46d1145cb78416d WatchSource:0}: Error finding container 151eab162fb07c7bec047f0b58f05784e69cfb48ede30f78e46d1145cb78416d: Status 404 returned error can't find the container with id 151eab162fb07c7bec047f0b58f05784e69cfb48ede30f78e46d1145cb78416d Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.649419 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.659355 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl"] Sep 30 08:17:47 crc kubenswrapper[4810]: W0930 08:17:47.674437 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11f06c30_1cef_40f3_a493_8be235930b71.slice/crio-b49daecc233f7426fca7a7a64ffe41e2848d5c5c87c6fdda407d43a4bf1eb29f WatchSource:0}: Error finding container b49daecc233f7426fca7a7a64ffe41e2848d5c5c87c6fdda407d43a4bf1eb29f: Status 404 returned error can't find the container with id b49daecc233f7426fca7a7a64ffe41e2848d5c5c87c6fdda407d43a4bf1eb29f Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.680775 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.846127 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4e9c80fa-52fc-41db-a0be-babb0bd14b56-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-9gsjq\" (UID: \"4e9c80fa-52fc-41db-a0be-babb0bd14b56\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.858376 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.867177 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4e9c80fa-52fc-41db-a0be-babb0bd14b56-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-9gsjq\" (UID: \"4e9c80fa-52fc-41db-a0be-babb0bd14b56\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.872200 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-wnngc"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.890792 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7"] Sep 30 08:17:47 crc kubenswrapper[4810]: E0930 08:17:47.897836 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g4dt4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-b8d54b5d7-g8ls7_openstack-operators(0cc20b2d-4504-4f46-b0a0-5aee4b2201a3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.914623 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc"] Sep 30 08:17:47 crc kubenswrapper[4810]: E0930 08:17:47.930022 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qmqgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-bc7dc7bd9-tc2wc_openstack-operators(3bbbc1cd-370d-468f-8985-cdf9219ca326): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 08:17:47 crc kubenswrapper[4810]: E0930 08:17:47.930195 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:de99ad053f95f132f62b38335b2e8bf22fc28acbd441c3814764d63b63ef755f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pmljr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-7d857cc749-v9pjf_openstack-operators(8b3a8007-4969-4c86-a03a-ce633b3c31aa): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.932863 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.936904 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4"] Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.941259 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-g8s55"] Sep 30 08:17:47 crc kubenswrapper[4810]: E0930 08:17:47.949525 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:e6f1ed6b386f77415c2a44e770d98ab6d16b6f6b494c4d1b4ac4b46368c4a4e6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8jmkb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-7975b88857-hd4x4_openstack-operators(8c5f7678-0f63-4323-910c-227040fdac66): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 08:17:47 crc kubenswrapper[4810]: E0930 08:17:47.958877 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6tbfw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-f66b554c6-g8s55_openstack-operators(7fe8706c-6dfa-4a2f-876b-4cded24209bf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 08:17:47 crc kubenswrapper[4810]: I0930 08:17:47.997670 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.050729 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958"] Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.056096 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw"] Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.068517 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x"] Sep 30 08:17:48 crc kubenswrapper[4810]: W0930 08:17:48.080235 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb63222b9_1cc3_4a55_a498_ce538cea80b9.slice/crio-36012fa0cc38a348a73c3daa5613f4e59dfd11b5d6be13eb4b4f59c508f4dd6a WatchSource:0}: Error finding container 36012fa0cc38a348a73c3daa5613f4e59dfd11b5d6be13eb4b4f59c508f4dd6a: Status 404 returned error can't find the container with id 36012fa0cc38a348a73c3daa5613f4e59dfd11b5d6be13eb4b4f59c508f4dd6a Sep 30 08:17:48 crc kubenswrapper[4810]: W0930 08:17:48.092054 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf4551e1_3981_407f_ba07_bff4db2b6d6d.slice/crio-ec2550b0abbd0c6257580712e18c5e12710b8b75af6eee88f10ac8f26d5c11ed WatchSource:0}: Error finding container ec2550b0abbd0c6257580712e18c5e12710b8b75af6eee88f10ac8f26d5c11ed: Status 404 returned error can't find the container with id ec2550b0abbd0c6257580712e18c5e12710b8b75af6eee88f10ac8f26d5c11ed Sep 30 08:17:48 crc kubenswrapper[4810]: W0930 08:17:48.094411 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3df3a4ac_fc49_404f_aaba_cc8c053365c4.slice/crio-432c89fbafab82fe8d6b8d3396b040f2c4f1ff4b766d2ea356189f7225d07c79 WatchSource:0}: Error finding container 432c89fbafab82fe8d6b8d3396b040f2c4f1ff4b766d2ea356189f7225d07c79: Status 404 returned error can't find the container with id 432c89fbafab82fe8d6b8d3396b040f2c4f1ff4b766d2ea356189f7225d07c79 Sep 30 08:17:48 crc kubenswrapper[4810]: E0930 08:17:48.096213 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9vmjr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-79d8469568-js8pw_openstack-operators(bf4551e1-3981-407f-ba07-bff4db2b6d6d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 08:17:48 crc kubenswrapper[4810]: E0930 08:17:48.097413 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" podUID="bf4551e1-3981-407f-ba07-bff4db2b6d6d" Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.431863 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf" event={"ID":"e801f35d-5d9f-48df-9a66-dd11ca14d26c","Type":"ContainerStarted","Data":"c6d5d2c731fab3573bc748fd21228c163ca41acb643f7cebb7dcbffcc493d00d"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.432999 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt" event={"ID":"b32751fb-95c0-4295-937a-ce7f82562863","Type":"ContainerStarted","Data":"60e24dbe5fe5586e93fccf47b533f326d700ba719ac2ca7231d643a5209eb029"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.434019 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" event={"ID":"bf4551e1-3981-407f-ba07-bff4db2b6d6d","Type":"ContainerStarted","Data":"ec2550b0abbd0c6257580712e18c5e12710b8b75af6eee88f10ac8f26d5c11ed"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.438924 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" event={"ID":"8c5f7678-0f63-4323-910c-227040fdac66","Type":"ContainerStarted","Data":"d975df601c522e4a75b766462394e379bde68e93d83ba26784a1133a6462b3ad"} Sep 30 08:17:48 crc kubenswrapper[4810]: E0930 08:17:48.438938 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" podUID="bf4551e1-3981-407f-ba07-bff4db2b6d6d" Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.447956 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd" event={"ID":"52a2d49f-0736-4cac-8363-1a71fa6c1dd4","Type":"ContainerStarted","Data":"f80c10d03fb2ea58a22b58db326f3540c71120c02360ffce1e44ac0b6676cb23"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.466139 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd" event={"ID":"df615184-edc5-4d93-b0ae-b414d9eb0ce1","Type":"ContainerStarted","Data":"7f70cc06e40267162f23594ea4380ec174af82ea910d71c7bc625f418824cc16"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.467976 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl" event={"ID":"11f06c30-1cef-40f3-a493-8be235930b71","Type":"ContainerStarted","Data":"b49daecc233f7426fca7a7a64ffe41e2848d5c5c87c6fdda407d43a4bf1eb29f"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.471124 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-wnngc" event={"ID":"8a8a969d-2c5d-4229-b01d-cabcf3aec35c","Type":"ContainerStarted","Data":"22b8f6aceaf6f02771224b21dbeaf4d6fb253f51b31c2bce66cf5e726bd96783"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.475904 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm" event={"ID":"0f34a7d1-aa5e-4b55-a75d-5366e196e742","Type":"ContainerStarted","Data":"0b81adac2aad638df994065a7518eaeab52880b0e35ce41c96bf7da33dded824"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.477233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" event={"ID":"8b3a8007-4969-4c86-a03a-ce633b3c31aa","Type":"ContainerStarted","Data":"e3c7d06b7079336a77dba60cf235eb9b61cf85b9f28587961db40956cab6141d"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.481204 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958" event={"ID":"b63222b9-1cc3-4a55-a498-ce538cea80b9","Type":"ContainerStarted","Data":"36012fa0cc38a348a73c3daa5613f4e59dfd11b5d6be13eb4b4f59c508f4dd6a"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.482893 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" event={"ID":"0cc20b2d-4504-4f46-b0a0-5aee4b2201a3","Type":"ContainerStarted","Data":"7e72d0305ed2e5ef65c534fd5bba1f3f5cd96d0330be75a326b1dce6f74a6e3b"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.491408 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" event={"ID":"7fe8706c-6dfa-4a2f-876b-4cded24209bf","Type":"ContainerStarted","Data":"4b7a49c092ee066961bbd5e73ea4570c7d85a210a234e53dc6bb310cba0d7980"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.503456 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j" event={"ID":"24044399-d033-436a-ad32-1e1aefe76626","Type":"ContainerStarted","Data":"151eab162fb07c7bec047f0b58f05784e69cfb48ede30f78e46d1145cb78416d"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.509339 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" event={"ID":"3bbbc1cd-370d-468f-8985-cdf9219ca326","Type":"ContainerStarted","Data":"4cee94a8724a9c43a98d446c5fcf3f7117dc0f8144dd460c3a2a3e3f0a603d6e"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.525473 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" event={"ID":"3df3a4ac-fc49-404f-aaba-cc8c053365c4","Type":"ContainerStarted","Data":"432c89fbafab82fe8d6b8d3396b040f2c4f1ff4b766d2ea356189f7225d07c79"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.542714 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm" event={"ID":"ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e","Type":"ContainerStarted","Data":"d038891fe2a1680ff4f810a0b2409fe1945e4030ade67981817499d32b31256b"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.550947 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn" event={"ID":"02d77694-4bb0-4735-94c1-a3314ee634b9","Type":"ContainerStarted","Data":"83e37d64819baa8ce5b4892798b29e90f97e8dd12127e9ad9238310d66ae4ded"} Sep 30 08:17:48 crc kubenswrapper[4810]: I0930 08:17:48.607167 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq"] Sep 30 08:17:48 crc kubenswrapper[4810]: W0930 08:17:48.617500 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e9c80fa_52fc_41db_a0be_babb0bd14b56.slice/crio-7ddec7ed7c82d6fbf6ae8157bf4116641efb009e71e26f6b8f6477d9923fa33e WatchSource:0}: Error finding container 7ddec7ed7c82d6fbf6ae8157bf4116641efb009e71e26f6b8f6477d9923fa33e: Status 404 returned error can't find the container with id 7ddec7ed7c82d6fbf6ae8157bf4116641efb009e71e26f6b8f6477d9923fa33e Sep 30 08:17:49 crc kubenswrapper[4810]: E0930 08:17:49.364878 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" podUID="8c5f7678-0f63-4323-910c-227040fdac66" Sep 30 08:17:49 crc kubenswrapper[4810]: E0930 08:17:49.415999 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" podUID="3bbbc1cd-370d-468f-8985-cdf9219ca326" Sep 30 08:17:49 crc kubenswrapper[4810]: E0930 08:17:49.453576 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" podUID="8b3a8007-4969-4c86-a03a-ce633b3c31aa" Sep 30 08:17:49 crc kubenswrapper[4810]: E0930 08:17:49.453732 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" podUID="7fe8706c-6dfa-4a2f-876b-4cded24209bf" Sep 30 08:17:49 crc kubenswrapper[4810]: E0930 08:17:49.453879 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" podUID="0cc20b2d-4504-4f46-b0a0-5aee4b2201a3" Sep 30 08:17:49 crc kubenswrapper[4810]: I0930 08:17:49.584759 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" event={"ID":"3df3a4ac-fc49-404f-aaba-cc8c053365c4","Type":"ContainerStarted","Data":"c1a4b2851dc870c1eb1a69f397f9ee409d4e6dd144097983af5e55c45e774a3c"} Sep 30 08:17:49 crc kubenswrapper[4810]: I0930 08:17:49.586605 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" event={"ID":"8b3a8007-4969-4c86-a03a-ce633b3c31aa","Type":"ContainerStarted","Data":"ffd8c60c1b72a050087b658f396726b824ffbeae79ff1f2f1fc03e13c66f47f8"} Sep 30 08:17:49 crc kubenswrapper[4810]: I0930 08:17:49.587950 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" event={"ID":"8c5f7678-0f63-4323-910c-227040fdac66","Type":"ContainerStarted","Data":"69cbfeeba47266dc3ca63d8ff0210635f4b0a78c57cc6d5bcf5c45cec674cf83"} Sep 30 08:17:49 crc kubenswrapper[4810]: E0930 08:17:49.588751 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:de99ad053f95f132f62b38335b2e8bf22fc28acbd441c3814764d63b63ef755f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" podUID="8b3a8007-4969-4c86-a03a-ce633b3c31aa" Sep 30 08:17:49 crc kubenswrapper[4810]: E0930 08:17:49.589822 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:e6f1ed6b386f77415c2a44e770d98ab6d16b6f6b494c4d1b4ac4b46368c4a4e6\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" podUID="8c5f7678-0f63-4323-910c-227040fdac66" Sep 30 08:17:49 crc kubenswrapper[4810]: I0930 08:17:49.595482 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" event={"ID":"4e9c80fa-52fc-41db-a0be-babb0bd14b56","Type":"ContainerStarted","Data":"7ddec7ed7c82d6fbf6ae8157bf4116641efb009e71e26f6b8f6477d9923fa33e"} Sep 30 08:17:49 crc kubenswrapper[4810]: I0930 08:17:49.598664 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" event={"ID":"0cc20b2d-4504-4f46-b0a0-5aee4b2201a3","Type":"ContainerStarted","Data":"5093567ab7a28e2ca21c2001b9c3f3a37dcda9b5c7e997ec618ea30b685cc4b3"} Sep 30 08:17:49 crc kubenswrapper[4810]: E0930 08:17:49.603425 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" podUID="0cc20b2d-4504-4f46-b0a0-5aee4b2201a3" Sep 30 08:17:49 crc kubenswrapper[4810]: I0930 08:17:49.622327 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" event={"ID":"7fe8706c-6dfa-4a2f-876b-4cded24209bf","Type":"ContainerStarted","Data":"38cc7d5799c7eddd2a678e87af4c39ea55d67700a15e4413d4b8f5b7a1f8e7a5"} Sep 30 08:17:49 crc kubenswrapper[4810]: E0930 08:17:49.625622 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" podUID="7fe8706c-6dfa-4a2f-876b-4cded24209bf" Sep 30 08:17:49 crc kubenswrapper[4810]: I0930 08:17:49.648387 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" event={"ID":"3bbbc1cd-370d-468f-8985-cdf9219ca326","Type":"ContainerStarted","Data":"1c54da0efdd5d80d193a5a73518936ebc63a9d51c1a4834bb4b5931c3a0b0e28"} Sep 30 08:17:49 crc kubenswrapper[4810]: E0930 08:17:49.652123 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" podUID="bf4551e1-3981-407f-ba07-bff4db2b6d6d" Sep 30 08:17:49 crc kubenswrapper[4810]: E0930 08:17:49.652217 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" podUID="3bbbc1cd-370d-468f-8985-cdf9219ca326" Sep 30 08:17:50 crc kubenswrapper[4810]: I0930 08:17:50.672277 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" event={"ID":"3df3a4ac-fc49-404f-aaba-cc8c053365c4","Type":"ContainerStarted","Data":"e810d4ce3950a74e6bf362f51abd5b937b4dca1408fe46b791a9a1ad66086d5f"} Sep 30 08:17:50 crc kubenswrapper[4810]: I0930 08:17:50.672980 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" Sep 30 08:17:50 crc kubenswrapper[4810]: E0930 08:17:50.674925 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" podUID="0cc20b2d-4504-4f46-b0a0-5aee4b2201a3" Sep 30 08:17:50 crc kubenswrapper[4810]: E0930 08:17:50.674925 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:e6f1ed6b386f77415c2a44e770d98ab6d16b6f6b494c4d1b4ac4b46368c4a4e6\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" podUID="8c5f7678-0f63-4323-910c-227040fdac66" Sep 30 08:17:50 crc kubenswrapper[4810]: E0930 08:17:50.674962 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" podUID="7fe8706c-6dfa-4a2f-876b-4cded24209bf" Sep 30 08:17:50 crc kubenswrapper[4810]: E0930 08:17:50.674983 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:de99ad053f95f132f62b38335b2e8bf22fc28acbd441c3814764d63b63ef755f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" podUID="8b3a8007-4969-4c86-a03a-ce633b3c31aa" Sep 30 08:17:50 crc kubenswrapper[4810]: E0930 08:17:50.676948 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" podUID="3bbbc1cd-370d-468f-8985-cdf9219ca326" Sep 30 08:17:50 crc kubenswrapper[4810]: I0930 08:17:50.726106 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" podStartSLOduration=4.726083569 podStartE2EDuration="4.726083569s" podCreationTimestamp="2025-09-30 08:17:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:17:50.718111835 +0000 UTC m=+894.170311122" watchObservedRunningTime="2025-09-30 08:17:50.726083569 +0000 UTC m=+894.178282836" Sep 30 08:17:56 crc kubenswrapper[4810]: I0930 08:17:56.960297 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-64647d55fc-6z69x" Sep 30 08:18:03 crc kubenswrapper[4810]: I0930 08:18:03.791699 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" event={"ID":"4e9c80fa-52fc-41db-a0be-babb0bd14b56","Type":"ContainerStarted","Data":"1c0b79b013b3311d2c7bd375a67b4d9ea53a7ac70e54f1eca31d35c63af5ff08"} Sep 30 08:18:03 crc kubenswrapper[4810]: I0930 08:18:03.805987 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf" event={"ID":"e801f35d-5d9f-48df-9a66-dd11ca14d26c","Type":"ContainerStarted","Data":"56d117548b547993badd6f69a71698d46f9c46d8c3a16e95e9041c63dff4e892"} Sep 30 08:18:03 crc kubenswrapper[4810]: I0930 08:18:03.819860 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw" event={"ID":"09ab9971-634a-40dd-88df-c1093fb66766","Type":"ContainerStarted","Data":"e213e6e03ccc39f8b114955f8338d2bacf211bec73fd271069adb637ec8c2d51"} Sep 30 08:18:03 crc kubenswrapper[4810]: I0930 08:18:03.821282 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl" event={"ID":"11f06c30-1cef-40f3-a493-8be235930b71","Type":"ContainerStarted","Data":"46688dd21539c66c08539e15ca3d2ce93c34a51251b282aa8b4090f312168747"} Sep 30 08:18:03 crc kubenswrapper[4810]: I0930 08:18:03.825638 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m" event={"ID":"6ad79408-8dc2-4e7e-aee3-54e3778a344e","Type":"ContainerStarted","Data":"50c367b385fb4c7d630c1a3c3d4ff097563961ec55d3d83c827b907f4570069e"} Sep 30 08:18:03 crc kubenswrapper[4810]: I0930 08:18:03.841900 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958" event={"ID":"b63222b9-1cc3-4a55-a498-ce538cea80b9","Type":"ContainerStarted","Data":"802f1f79d79f31102e504d5942fe13c57c1a22f3bda62ea2b91b20c4fb23fc8d"} Sep 30 08:18:03 crc kubenswrapper[4810]: I0930 08:18:03.844691 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd" event={"ID":"52a2d49f-0736-4cac-8363-1a71fa6c1dd4","Type":"ContainerStarted","Data":"d055f398ba56e5f1ce23867082aac0c9e3890ce52d79e37b50ee49e84093b6af"} Sep 30 08:18:03 crc kubenswrapper[4810]: I0930 08:18:03.849093 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs" event={"ID":"da3d67ab-80be-486a-bc4d-d36ae6eb4756","Type":"ContainerStarted","Data":"29a6824163c184ef77e426a834fadcf69aada64a5c3bc13f860ea3279bd4415c"} Sep 30 08:18:03 crc kubenswrapper[4810]: I0930 08:18:03.853699 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j" event={"ID":"24044399-d033-436a-ad32-1e1aefe76626","Type":"ContainerStarted","Data":"ab5135cb3607cf706808ad2619d712749c8aa2b16faf651fc3be24df999987b3"} Sep 30 08:18:03 crc kubenswrapper[4810]: I0930 08:18:03.860550 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-wnngc" event={"ID":"8a8a969d-2c5d-4229-b01d-cabcf3aec35c","Type":"ContainerStarted","Data":"3a2b9ad188a8f16cecb25636cb2d34f4c251995c80b5cb706968f6ef98de352b"} Sep 30 08:18:03 crc kubenswrapper[4810]: I0930 08:18:03.897785 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8" event={"ID":"982a5d6f-5897-40d0-88a4-40a9f1ae9a5f","Type":"ContainerStarted","Data":"cac4c4c5c0e4d43f86e034bd304f660ea7952f8c2e2b291ceece34d9f83b621f"} Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.913626 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm" event={"ID":"0f34a7d1-aa5e-4b55-a75d-5366e196e742","Type":"ContainerStarted","Data":"e6de5314a9610b1b04dea435966417d4dfaf27072575977888dce700742beb53"} Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.913681 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm" event={"ID":"0f34a7d1-aa5e-4b55-a75d-5366e196e742","Type":"ContainerStarted","Data":"4264d3fe46ef84ae7263c2108ccd4e3b7f56bab187792806fd1e735527ec7fdd"} Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.913860 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm" Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.937989 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm" podStartSLOduration=5.125460165 podStartE2EDuration="19.937971478s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.887084076 +0000 UTC m=+891.339283343" lastFinishedPulling="2025-09-30 08:18:02.699595389 +0000 UTC m=+906.151794656" observedRunningTime="2025-09-30 08:18:04.935013031 +0000 UTC m=+908.387212298" watchObservedRunningTime="2025-09-30 08:18:04.937971478 +0000 UTC m=+908.390170745" Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.950681 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd" event={"ID":"52a2d49f-0736-4cac-8363-1a71fa6c1dd4","Type":"ContainerStarted","Data":"ace53712e172f97f52e41e912a68ff9f392f51801bd70d95ec6964882b71cce3"} Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.950756 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd" Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.962957 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs" event={"ID":"da3d67ab-80be-486a-bc4d-d36ae6eb4756","Type":"ContainerStarted","Data":"fde69c0ab6c0d3714500a63e8075afd3fa38acb9176b0828ec1301434a23abcc"} Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.963023 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs" Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.965853 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw" event={"ID":"09ab9971-634a-40dd-88df-c1093fb66766","Type":"ContainerStarted","Data":"a8abad1cacf9286ebfce61a9e6af5adec91212c0df87e8756a561ebfa64a5dc0"} Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.966678 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw" Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.967076 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd" podStartSLOduration=4.714413933 podStartE2EDuration="19.967059761s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.447280891 +0000 UTC m=+890.899480158" lastFinishedPulling="2025-09-30 08:18:02.699926729 +0000 UTC m=+906.152125986" observedRunningTime="2025-09-30 08:18:04.9635953 +0000 UTC m=+908.415794567" watchObservedRunningTime="2025-09-30 08:18:04.967059761 +0000 UTC m=+908.419259028" Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.969077 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn" event={"ID":"02d77694-4bb0-4735-94c1-a3314ee634b9","Type":"ContainerStarted","Data":"076da4157d6870c05192a123b531b1567821715e2dc7849297c46ab0fc9d3670"} Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.972294 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" event={"ID":"4e9c80fa-52fc-41db-a0be-babb0bd14b56","Type":"ContainerStarted","Data":"cb4f95e1e97d3691428ac68854afae93f4d2b6edf23072c451aad42c8edcf372"} Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.972567 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.974792 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm" event={"ID":"ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e","Type":"ContainerStarted","Data":"ce2a34101c5e8f197fb51aef461de653c3ac9355d06c95da9c85da4dba2d2cf3"} Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.976293 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8" event={"ID":"982a5d6f-5897-40d0-88a4-40a9f1ae9a5f","Type":"ContainerStarted","Data":"bb837aacb914c0c3db7701fb51e1fe5e2b91eba38c9bad237bebeb510ad08130"} Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.976489 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8" Sep 30 08:18:04 crc kubenswrapper[4810]: I0930 08:18:04.988549 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs" podStartSLOduration=4.562252109 podStartE2EDuration="19.988529751s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.381903233 +0000 UTC m=+890.834102500" lastFinishedPulling="2025-09-30 08:18:02.808180875 +0000 UTC m=+906.260380142" observedRunningTime="2025-09-30 08:18:04.983762722 +0000 UTC m=+908.435961989" watchObservedRunningTime="2025-09-30 08:18:04.988529751 +0000 UTC m=+908.440729008" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:04.997822 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-wnngc" event={"ID":"8a8a969d-2c5d-4229-b01d-cabcf3aec35c","Type":"ContainerStarted","Data":"8fca94c86d20010fee1e44750c315d9e0309e32cf043b156991fe65ba2670d30"} Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:04.998571 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-wnngc" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.027293 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m" event={"ID":"6ad79408-8dc2-4e7e-aee3-54e3778a344e","Type":"ContainerStarted","Data":"4668649fd6b91fb809ae93f53204d13341226784ec580a5126d99e2d49f070cf"} Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.027967 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.028948 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8" podStartSLOduration=4.494443299 podStartE2EDuration="20.028936357s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.076889313 +0000 UTC m=+890.529088580" lastFinishedPulling="2025-09-30 08:18:02.611382341 +0000 UTC m=+906.063581638" observedRunningTime="2025-09-30 08:18:05.017928514 +0000 UTC m=+908.470127781" watchObservedRunningTime="2025-09-30 08:18:05.028936357 +0000 UTC m=+908.481135624" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.050567 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw" podStartSLOduration=4.840221175 podStartE2EDuration="20.050547681s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.381853641 +0000 UTC m=+890.834052898" lastFinishedPulling="2025-09-30 08:18:02.592180097 +0000 UTC m=+906.044379404" observedRunningTime="2025-09-30 08:18:05.036368855 +0000 UTC m=+908.488568122" watchObservedRunningTime="2025-09-30 08:18:05.050547681 +0000 UTC m=+908.502746948" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.057456 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf" event={"ID":"e801f35d-5d9f-48df-9a66-dd11ca14d26c","Type":"ContainerStarted","Data":"076fa97e1256bb376fb97dd87248dfeee3f27baa981a0e481ea8ce4af1d14abe"} Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.058136 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.061574 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j" event={"ID":"24044399-d033-436a-ad32-1e1aefe76626","Type":"ContainerStarted","Data":"e9607f3a924a293683a8c1de336d41007ac5d3ed38bfeca26728cdda09a2dbe9"} Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.062183 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.064698 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl" event={"ID":"11f06c30-1cef-40f3-a493-8be235930b71","Type":"ContainerStarted","Data":"361dbfd746a359539e671f41eac0d736e1c06fcdeda98a5567eba11f8f93def3"} Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.065022 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.066806 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958" event={"ID":"b63222b9-1cc3-4a55-a498-ce538cea80b9","Type":"ContainerStarted","Data":"14fcf67727ee14fb514d05977e5560a59c86bb37d6299f4ecd14e4df308abc91"} Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.066991 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.068894 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd" event={"ID":"df615184-edc5-4d93-b0ae-b414d9eb0ce1","Type":"ContainerStarted","Data":"d952a6c5b1ed31f0a18cb1d0022ca4cf4f62ecb66c2557a9214b03c39355c310"} Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.068938 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd" event={"ID":"df615184-edc5-4d93-b0ae-b414d9eb0ce1","Type":"ContainerStarted","Data":"ca2c7554a0f26e2c7301ba962a7ead0df7373c91d8227dff493fe56efa110cab"} Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.069123 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.070507 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt" event={"ID":"b32751fb-95c0-4295-937a-ce7f82562863","Type":"ContainerStarted","Data":"609ef5eb77a8920fe91e70a4ffb78deb35aa9a8694acea95041913ba4fdf2f25"} Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.104574 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j" podStartSLOduration=5.65457165 podStartE2EDuration="20.104559516s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.652823842 +0000 UTC m=+891.105023109" lastFinishedPulling="2025-09-30 08:18:02.102811708 +0000 UTC m=+905.555010975" observedRunningTime="2025-09-30 08:18:05.102190117 +0000 UTC m=+908.554389374" watchObservedRunningTime="2025-09-30 08:18:05.104559516 +0000 UTC m=+908.556758783" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.106200 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" podStartSLOduration=5.965190254 podStartE2EDuration="20.106195554s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:48.630037836 +0000 UTC m=+892.082237103" lastFinishedPulling="2025-09-30 08:18:02.771043136 +0000 UTC m=+906.223242403" observedRunningTime="2025-09-30 08:18:05.077640616 +0000 UTC m=+908.529839903" watchObservedRunningTime="2025-09-30 08:18:05.106195554 +0000 UTC m=+908.558394821" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.134952 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf" podStartSLOduration=4.698065273 podStartE2EDuration="20.134932317s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.439885144 +0000 UTC m=+890.892084411" lastFinishedPulling="2025-09-30 08:18:02.876752188 +0000 UTC m=+906.328951455" observedRunningTime="2025-09-30 08:18:05.131708893 +0000 UTC m=+908.583908160" watchObservedRunningTime="2025-09-30 08:18:05.134932317 +0000 UTC m=+908.587131584" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.158559 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m" podStartSLOduration=6.271143262 podStartE2EDuration="20.15854508s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:46.930071265 +0000 UTC m=+890.382270532" lastFinishedPulling="2025-09-30 08:18:00.817473083 +0000 UTC m=+904.269672350" observedRunningTime="2025-09-30 08:18:05.152327308 +0000 UTC m=+908.604526575" watchObservedRunningTime="2025-09-30 08:18:05.15854508 +0000 UTC m=+908.610744347" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.174888 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-wnngc" podStartSLOduration=5.463674459 podStartE2EDuration="20.174778527s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.878934127 +0000 UTC m=+891.331133394" lastFinishedPulling="2025-09-30 08:18:02.590038165 +0000 UTC m=+906.042237462" observedRunningTime="2025-09-30 08:18:05.17013265 +0000 UTC m=+908.622331917" watchObservedRunningTime="2025-09-30 08:18:05.174778527 +0000 UTC m=+908.626977794" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.193337 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958" podStartSLOduration=4.689573288 podStartE2EDuration="19.193320401s" podCreationTimestamp="2025-09-30 08:17:46 +0000 UTC" firstStartedPulling="2025-09-30 08:17:48.088007772 +0000 UTC m=+891.540207039" lastFinishedPulling="2025-09-30 08:18:02.591754855 +0000 UTC m=+906.043954152" observedRunningTime="2025-09-30 08:18:05.186639375 +0000 UTC m=+908.638838642" watchObservedRunningTime="2025-09-30 08:18:05.193320401 +0000 UTC m=+908.645519668" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.205618 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd" podStartSLOduration=4.883078913 podStartE2EDuration="20.205593341s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.451567047 +0000 UTC m=+890.903766314" lastFinishedPulling="2025-09-30 08:18:02.774081475 +0000 UTC m=+906.226280742" observedRunningTime="2025-09-30 08:18:05.203564181 +0000 UTC m=+908.655763448" watchObservedRunningTime="2025-09-30 08:18:05.205593341 +0000 UTC m=+908.657792608" Sep 30 08:18:05 crc kubenswrapper[4810]: I0930 08:18:05.227657 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl" podStartSLOduration=5.879783728 podStartE2EDuration="20.227642038s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.67764043 +0000 UTC m=+891.129839697" lastFinishedPulling="2025-09-30 08:18:02.02549875 +0000 UTC m=+905.477698007" observedRunningTime="2025-09-30 08:18:05.222626071 +0000 UTC m=+908.674825338" watchObservedRunningTime="2025-09-30 08:18:05.227642038 +0000 UTC m=+908.679841305" Sep 30 08:18:08 crc kubenswrapper[4810]: I0930 08:18:08.016357 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-9gsjq" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.169953 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" event={"ID":"8c5f7678-0f63-4323-910c-227040fdac66","Type":"ContainerStarted","Data":"42d593bc2cf5a9c61fbc2180cd0b5de07b53c900932079fe096a015d62344449"} Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.171400 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.173236 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" event={"ID":"3bbbc1cd-370d-468f-8985-cdf9219ca326","Type":"ContainerStarted","Data":"b591421e50ecd37687397cc895f114ccc6783312190fdfc27440187cde0f02c0"} Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.173426 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.175710 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" event={"ID":"7fe8706c-6dfa-4a2f-876b-4cded24209bf","Type":"ContainerStarted","Data":"11ec001caf05b4cd94e8a2f945084f1ca9c9855a7cdad17e93792bb219d23824"} Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.176149 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.178256 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt" event={"ID":"b32751fb-95c0-4295-937a-ce7f82562863","Type":"ContainerStarted","Data":"7aac2fbd5bf976db6ce887c086dbaf6f1ff09c5fbc03125ea2c56cfc3c86769e"} Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.178429 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.181147 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.182966 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm" event={"ID":"ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e","Type":"ContainerStarted","Data":"0343389947d6dbad0d8f1341d544d028f31e71c0022c0621f3e9dde3387a0d8b"} Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.184393 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.186123 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.200835 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn" event={"ID":"02d77694-4bb0-4735-94c1-a3314ee634b9","Type":"ContainerStarted","Data":"9fd919e011c110ea88990992763592d25213991a0c5d8ec2e933a49caba0e3a8"} Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.200959 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.204080 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" podStartSLOduration=3.711249319 podStartE2EDuration="26.20404935s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.94925229 +0000 UTC m=+891.401451557" lastFinishedPulling="2025-09-30 08:18:10.442052321 +0000 UTC m=+913.894251588" observedRunningTime="2025-09-30 08:18:11.197736145 +0000 UTC m=+914.649935412" watchObservedRunningTime="2025-09-30 08:18:11.20404935 +0000 UTC m=+914.656248617" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.206106 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.210446 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" event={"ID":"8b3a8007-4969-4c86-a03a-ce633b3c31aa","Type":"ContainerStarted","Data":"50f5059525511426dcecbe2a967d5f63f158669d76409a23f412ea21147cab08"} Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.211359 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.233794 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-v58wt" podStartSLOduration=11.153891093 podStartE2EDuration="26.233778162s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.691982251 +0000 UTC m=+891.144181518" lastFinishedPulling="2025-09-30 08:18:02.77186932 +0000 UTC m=+906.224068587" observedRunningTime="2025-09-30 08:18:11.231098533 +0000 UTC m=+914.683297800" watchObservedRunningTime="2025-09-30 08:18:11.233778162 +0000 UTC m=+914.685977429" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.255842 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" podStartSLOduration=2.789355481 podStartE2EDuration="25.255814209s" podCreationTimestamp="2025-09-30 08:17:46 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.958549613 +0000 UTC m=+891.410748880" lastFinishedPulling="2025-09-30 08:18:10.425008341 +0000 UTC m=+913.877207608" observedRunningTime="2025-09-30 08:18:11.25312959 +0000 UTC m=+914.705328857" watchObservedRunningTime="2025-09-30 08:18:11.255814209 +0000 UTC m=+914.708013476" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.281361 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" podStartSLOduration=2.53370637 podStartE2EDuration="25.281328827s" podCreationTimestamp="2025-09-30 08:17:46 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.92980508 +0000 UTC m=+891.382004347" lastFinishedPulling="2025-09-30 08:18:10.677427537 +0000 UTC m=+914.129626804" observedRunningTime="2025-09-30 08:18:11.277753632 +0000 UTC m=+914.729952899" watchObservedRunningTime="2025-09-30 08:18:11.281328827 +0000 UTC m=+914.733528094" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.302591 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-t6wwm" podStartSLOduration=11.388457556 podStartE2EDuration="26.30256437s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.657141499 +0000 UTC m=+891.109340766" lastFinishedPulling="2025-09-30 08:18:02.571248263 +0000 UTC m=+906.023447580" observedRunningTime="2025-09-30 08:18:11.296155382 +0000 UTC m=+914.748354649" watchObservedRunningTime="2025-09-30 08:18:11.30256437 +0000 UTC m=+914.754763637" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.358291 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" podStartSLOduration=3.846606881 podStartE2EDuration="26.358248664s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.930115379 +0000 UTC m=+891.382314646" lastFinishedPulling="2025-09-30 08:18:10.441757162 +0000 UTC m=+913.893956429" observedRunningTime="2025-09-30 08:18:11.347097927 +0000 UTC m=+914.799297204" watchObservedRunningTime="2025-09-30 08:18:11.358248664 +0000 UTC m=+914.810447931" Sep 30 08:18:11 crc kubenswrapper[4810]: I0930 08:18:11.379306 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-zfcrn" podStartSLOduration=11.203651644 podStartE2EDuration="26.379284742s" podCreationTimestamp="2025-09-30 08:17:45 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.418728333 +0000 UTC m=+890.870927600" lastFinishedPulling="2025-09-30 08:18:02.594361391 +0000 UTC m=+906.046560698" observedRunningTime="2025-09-30 08:18:11.366781085 +0000 UTC m=+914.818980352" watchObservedRunningTime="2025-09-30 08:18:11.379284742 +0000 UTC m=+914.831484009" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.000682 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-zzj8m" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.032244 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-6dhj8" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.099027 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-9wfgs" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.112326 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-npvdf" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.208382 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vz9pd" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.216352 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-8bwwd" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.217370 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-g76mw" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.326681 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-wnngc" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.454859 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-hd4x4" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.456439 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-sndbm" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.528591 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-s785j" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.529114 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-t24hl" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.697928 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-tc2wc" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.748711 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-v9pjf" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.784070 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-f66b554c6-g8s55" Sep 30 08:18:16 crc kubenswrapper[4810]: I0930 08:18:16.861495 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-59446d9d9-kb958" Sep 30 08:18:48 crc kubenswrapper[4810]: E0930 08:18:48.652576 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b" Sep 30 08:18:48 crc kubenswrapper[4810]: E0930 08:18:48.653818 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9vmjr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-79d8469568-js8pw_openstack-operators(bf4551e1-3981-407f-ba07-bff4db2b6d6d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 08:18:48 crc kubenswrapper[4810]: E0930 08:18:48.655397 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" podUID="bf4551e1-3981-407f-ba07-bff4db2b6d6d" Sep 30 08:18:49 crc kubenswrapper[4810]: I0930 08:18:49.638055 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" event={"ID":"0cc20b2d-4504-4f46-b0a0-5aee4b2201a3","Type":"ContainerStarted","Data":"8d567c58cb7505f7d17e5b4c4df998d1e612fc230f1e3846f0ff0a29a5561a36"} Sep 30 08:18:49 crc kubenswrapper[4810]: I0930 08:18:49.638572 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" Sep 30 08:18:49 crc kubenswrapper[4810]: I0930 08:18:49.661147 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" podStartSLOduration=2.984967881 podStartE2EDuration="1m3.661121843s" podCreationTimestamp="2025-09-30 08:17:46 +0000 UTC" firstStartedPulling="2025-09-30 08:17:47.897667807 +0000 UTC m=+891.349867074" lastFinishedPulling="2025-09-30 08:18:48.573821759 +0000 UTC m=+952.026021036" observedRunningTime="2025-09-30 08:18:49.656663422 +0000 UTC m=+953.108862689" watchObservedRunningTime="2025-09-30 08:18:49.661121843 +0000 UTC m=+953.113321110" Sep 30 08:18:56 crc kubenswrapper[4810]: I0930 08:18:56.729118 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-g8ls7" Sep 30 08:19:01 crc kubenswrapper[4810]: E0930 08:19:01.310260 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" podUID="bf4551e1-3981-407f-ba07-bff4db2b6d6d" Sep 30 08:19:13 crc kubenswrapper[4810]: I0930 08:19:13.896111 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" event={"ID":"bf4551e1-3981-407f-ba07-bff4db2b6d6d","Type":"ContainerStarted","Data":"2e8a70f95639c1b9f4cf9bd21557bb6943c5d2b96e1e4e9808314144340d662c"} Sep 30 08:19:13 crc kubenswrapper[4810]: I0930 08:19:13.921780 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-js8pw" podStartSLOduration=2.553937202 podStartE2EDuration="1m27.921756145s" podCreationTimestamp="2025-09-30 08:17:46 +0000 UTC" firstStartedPulling="2025-09-30 08:17:48.096109299 +0000 UTC m=+891.548308566" lastFinishedPulling="2025-09-30 08:19:13.463928242 +0000 UTC m=+976.916127509" observedRunningTime="2025-09-30 08:19:13.915113362 +0000 UTC m=+977.367312639" watchObservedRunningTime="2025-09-30 08:19:13.921756145 +0000 UTC m=+977.373955432" Sep 30 08:19:31 crc kubenswrapper[4810]: I0930 08:19:31.833179 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5547dd57d7-76dtr"] Sep 30 08:19:31 crc kubenswrapper[4810]: I0930 08:19:31.835304 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" Sep 30 08:19:31 crc kubenswrapper[4810]: I0930 08:19:31.839630 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-pk7m2" Sep 30 08:19:31 crc kubenswrapper[4810]: I0930 08:19:31.839980 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Sep 30 08:19:31 crc kubenswrapper[4810]: I0930 08:19:31.840151 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Sep 30 08:19:31 crc kubenswrapper[4810]: I0930 08:19:31.840337 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Sep 30 08:19:31 crc kubenswrapper[4810]: I0930 08:19:31.852740 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5547dd57d7-76dtr"] Sep 30 08:19:31 crc kubenswrapper[4810]: I0930 08:19:31.927407 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dc5bf6869-qgznr"] Sep 30 08:19:31 crc kubenswrapper[4810]: I0930 08:19:31.928821 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:19:31 crc kubenswrapper[4810]: I0930 08:19:31.934096 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Sep 30 08:19:31 crc kubenswrapper[4810]: I0930 08:19:31.953954 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dc5bf6869-qgznr"] Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.007012 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-dns-svc\") pod \"dnsmasq-dns-5dc5bf6869-qgznr\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.007091 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5lqz\" (UniqueName: \"kubernetes.io/projected/b5d305fa-0c3f-48b1-b17b-4f543398d53d-kube-api-access-h5lqz\") pod \"dnsmasq-dns-5547dd57d7-76dtr\" (UID: \"b5d305fa-0c3f-48b1-b17b-4f543398d53d\") " pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.007169 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d305fa-0c3f-48b1-b17b-4f543398d53d-config\") pod \"dnsmasq-dns-5547dd57d7-76dtr\" (UID: \"b5d305fa-0c3f-48b1-b17b-4f543398d53d\") " pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.007194 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-config\") pod \"dnsmasq-dns-5dc5bf6869-qgznr\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.007214 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bffj\" (UniqueName: \"kubernetes.io/projected/fd327763-4320-40ff-8d03-7f48ffbfdc1a-kube-api-access-5bffj\") pod \"dnsmasq-dns-5dc5bf6869-qgznr\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.108576 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5lqz\" (UniqueName: \"kubernetes.io/projected/b5d305fa-0c3f-48b1-b17b-4f543398d53d-kube-api-access-h5lqz\") pod \"dnsmasq-dns-5547dd57d7-76dtr\" (UID: \"b5d305fa-0c3f-48b1-b17b-4f543398d53d\") " pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.109008 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d305fa-0c3f-48b1-b17b-4f543398d53d-config\") pod \"dnsmasq-dns-5547dd57d7-76dtr\" (UID: \"b5d305fa-0c3f-48b1-b17b-4f543398d53d\") " pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.109101 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-config\") pod \"dnsmasq-dns-5dc5bf6869-qgznr\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.109173 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bffj\" (UniqueName: \"kubernetes.io/projected/fd327763-4320-40ff-8d03-7f48ffbfdc1a-kube-api-access-5bffj\") pod \"dnsmasq-dns-5dc5bf6869-qgznr\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.109314 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-dns-svc\") pod \"dnsmasq-dns-5dc5bf6869-qgznr\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.110086 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-dns-svc\") pod \"dnsmasq-dns-5dc5bf6869-qgznr\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.110730 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d305fa-0c3f-48b1-b17b-4f543398d53d-config\") pod \"dnsmasq-dns-5547dd57d7-76dtr\" (UID: \"b5d305fa-0c3f-48b1-b17b-4f543398d53d\") " pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.111290 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-config\") pod \"dnsmasq-dns-5dc5bf6869-qgznr\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.127923 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5lqz\" (UniqueName: \"kubernetes.io/projected/b5d305fa-0c3f-48b1-b17b-4f543398d53d-kube-api-access-h5lqz\") pod \"dnsmasq-dns-5547dd57d7-76dtr\" (UID: \"b5d305fa-0c3f-48b1-b17b-4f543398d53d\") " pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.139119 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bffj\" (UniqueName: \"kubernetes.io/projected/fd327763-4320-40ff-8d03-7f48ffbfdc1a-kube-api-access-5bffj\") pod \"dnsmasq-dns-5dc5bf6869-qgznr\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.171244 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.249686 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.511445 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5547dd57d7-76dtr"] Sep 30 08:19:32 crc kubenswrapper[4810]: I0930 08:19:32.572354 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dc5bf6869-qgznr"] Sep 30 08:19:33 crc kubenswrapper[4810]: I0930 08:19:33.113388 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" event={"ID":"b5d305fa-0c3f-48b1-b17b-4f543398d53d","Type":"ContainerStarted","Data":"b9e2c9e1dc3faad353fcdc7e88c85fea05807fcfbf1fdf63d8fc213f72a537dc"} Sep 30 08:19:33 crc kubenswrapper[4810]: I0930 08:19:33.115755 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" event={"ID":"fd327763-4320-40ff-8d03-7f48ffbfdc1a","Type":"ContainerStarted","Data":"56af85a2724b5e7077952fdc7107b1cdf8472635bc1a1ba6d4a67dcfa5b4af1d"} Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.607908 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5547dd57d7-76dtr"] Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.634659 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cccfc6d89-q2n82"] Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.635885 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.646156 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cccfc6d89-q2n82"] Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.782532 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-config\") pod \"dnsmasq-dns-5cccfc6d89-q2n82\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.782787 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xww7s\" (UniqueName: \"kubernetes.io/projected/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-kube-api-access-xww7s\") pod \"dnsmasq-dns-5cccfc6d89-q2n82\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.782829 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-dns-svc\") pod \"dnsmasq-dns-5cccfc6d89-q2n82\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.884701 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-config\") pod \"dnsmasq-dns-5cccfc6d89-q2n82\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.884790 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xww7s\" (UniqueName: \"kubernetes.io/projected/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-kube-api-access-xww7s\") pod \"dnsmasq-dns-5cccfc6d89-q2n82\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.884868 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-dns-svc\") pod \"dnsmasq-dns-5cccfc6d89-q2n82\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.885943 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-config\") pod \"dnsmasq-dns-5cccfc6d89-q2n82\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.886418 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-dns-svc\") pod \"dnsmasq-dns-5cccfc6d89-q2n82\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.924341 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xww7s\" (UniqueName: \"kubernetes.io/projected/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-kube-api-access-xww7s\") pod \"dnsmasq-dns-5cccfc6d89-q2n82\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.929048 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dc5bf6869-qgznr"] Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.955487 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.959158 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75f7846495-qsb5s"] Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.960516 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:19:35 crc kubenswrapper[4810]: I0930 08:19:35.969643 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75f7846495-qsb5s"] Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.090618 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-config\") pod \"dnsmasq-dns-75f7846495-qsb5s\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.090675 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6wt5\" (UniqueName: \"kubernetes.io/projected/9161c7fc-8a57-49cd-ae3d-7043933ecae6-kube-api-access-l6wt5\") pod \"dnsmasq-dns-75f7846495-qsb5s\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.090755 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-dns-svc\") pod \"dnsmasq-dns-75f7846495-qsb5s\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.195300 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-config\") pod \"dnsmasq-dns-75f7846495-qsb5s\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.195652 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6wt5\" (UniqueName: \"kubernetes.io/projected/9161c7fc-8a57-49cd-ae3d-7043933ecae6-kube-api-access-l6wt5\") pod \"dnsmasq-dns-75f7846495-qsb5s\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.195677 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-dns-svc\") pod \"dnsmasq-dns-75f7846495-qsb5s\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.196609 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-dns-svc\") pod \"dnsmasq-dns-75f7846495-qsb5s\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.197295 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-config\") pod \"dnsmasq-dns-75f7846495-qsb5s\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.256353 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6wt5\" (UniqueName: \"kubernetes.io/projected/9161c7fc-8a57-49cd-ae3d-7043933ecae6-kube-api-access-l6wt5\") pod \"dnsmasq-dns-75f7846495-qsb5s\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.348638 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cccfc6d89-q2n82"] Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.390690 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74ccbc479c-c75jf"] Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.391185 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.392835 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.426152 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74ccbc479c-c75jf"] Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.508949 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-dns-svc\") pod \"dnsmasq-dns-74ccbc479c-c75jf\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.509003 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-config\") pod \"dnsmasq-dns-74ccbc479c-c75jf\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.509032 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xkjn\" (UniqueName: \"kubernetes.io/projected/7601717f-f941-49aa-bb26-309cfeb9401d-kube-api-access-4xkjn\") pod \"dnsmasq-dns-74ccbc479c-c75jf\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.614399 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-dns-svc\") pod \"dnsmasq-dns-74ccbc479c-c75jf\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.614488 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-config\") pod \"dnsmasq-dns-74ccbc479c-c75jf\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.614548 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xkjn\" (UniqueName: \"kubernetes.io/projected/7601717f-f941-49aa-bb26-309cfeb9401d-kube-api-access-4xkjn\") pod \"dnsmasq-dns-74ccbc479c-c75jf\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.615600 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-dns-svc\") pod \"dnsmasq-dns-74ccbc479c-c75jf\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.616035 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-config\") pod \"dnsmasq-dns-74ccbc479c-c75jf\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.617602 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cccfc6d89-q2n82"] Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.649014 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xkjn\" (UniqueName: \"kubernetes.io/projected/7601717f-f941-49aa-bb26-309cfeb9401d-kube-api-access-4xkjn\") pod \"dnsmasq-dns-74ccbc479c-c75jf\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.751871 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.803961 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.805339 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.809236 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.809434 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6gwgr" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.809551 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.809651 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.826180 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.826509 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.832976 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.833119 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.920172 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.920299 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.920330 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.920361 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.920392 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.920421 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.920446 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5bb3b225-5221-417f-9914-737723ac52aa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.920471 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.920515 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-config-data\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.920547 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htf9r\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-kube-api-access-htf9r\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.920573 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5bb3b225-5221-417f-9914-737723ac52aa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:36 crc kubenswrapper[4810]: I0930 08:19:36.995531 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75f7846495-qsb5s"] Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.021877 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5bb3b225-5221-417f-9914-737723ac52aa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.021935 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.021980 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.021996 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.022019 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.022038 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.022056 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.022076 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5bb3b225-5221-417f-9914-737723ac52aa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.022089 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.022131 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-config-data\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.022154 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htf9r\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-kube-api-access-htf9r\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.023335 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.023602 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.023638 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.023973 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.026704 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-config-data\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.027311 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5bb3b225-5221-417f-9914-737723ac52aa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.029653 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.032227 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5bb3b225-5221-417f-9914-737723ac52aa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.032710 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.032846 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.041326 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htf9r\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-kube-api-access-htf9r\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.070574 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.123451 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.124953 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.128592 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.128591 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.128756 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.128977 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gx7b6" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.129076 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.129255 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.132864 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.136881 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.136913 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.170670 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" event={"ID":"9161c7fc-8a57-49cd-ae3d-7043933ecae6","Type":"ContainerStarted","Data":"61440a7aac7f390d75ca97d3995c538037dc758e60136ab5244f1d62752de9cb"} Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.172703 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" event={"ID":"b3ed1d67-4c86-49f1-8081-b89bc5d210d3","Type":"ContainerStarted","Data":"cebd1e83ac902bf75074f42db393fffe1274286690f1428739f176493d10a432"} Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.218033 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74ccbc479c-c75jf"] Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.224806 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j54sx\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-kube-api-access-j54sx\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.224851 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.224900 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.224919 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cf638129-9405-48c6-8458-143b297363fd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.224936 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cf638129-9405-48c6-8458-143b297363fd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.224958 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.224976 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.224992 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.225016 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.225553 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.225854 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: W0930 08:19:37.236409 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7601717f_f941_49aa_bb26_309cfeb9401d.slice/crio-8d8758a484d8872511786184de235b5de21f1b47ca1977cf385e3e7367cdc501 WatchSource:0}: Error finding container 8d8758a484d8872511786184de235b5de21f1b47ca1977cf385e3e7367cdc501: Status 404 returned error can't find the container with id 8d8758a484d8872511786184de235b5de21f1b47ca1977cf385e3e7367cdc501 Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.329716 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.329781 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.329828 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.329887 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j54sx\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-kube-api-access-j54sx\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.329936 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.330055 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.330083 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cf638129-9405-48c6-8458-143b297363fd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.330112 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cf638129-9405-48c6-8458-143b297363fd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.330149 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.330170 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.330191 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.331301 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.332829 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.333446 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.334356 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.334809 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.338653 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.341177 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.342751 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.356887 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cf638129-9405-48c6-8458-143b297363fd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.371683 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cf638129-9405-48c6-8458-143b297363fd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.375971 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j54sx\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-kube-api-access-j54sx\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.396117 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.462809 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.467475 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 08:19:37 crc kubenswrapper[4810]: W0930 08:19:37.473553 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bb3b225_5221_417f_9914_737723ac52aa.slice/crio-c3c7f0ab98f0a063109b6ecd0d4261b59e1118b8cc965b3ff513d4873de959b2 WatchSource:0}: Error finding container c3c7f0ab98f0a063109b6ecd0d4261b59e1118b8cc965b3ff513d4873de959b2: Status 404 returned error can't find the container with id c3c7f0ab98f0a063109b6ecd0d4261b59e1118b8cc965b3ff513d4873de959b2 Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.502934 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.504956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.507857 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-erlang-cookie" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.513903 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-server-conf" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.513920 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-config-data" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.516485 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-plugins-conf" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.516586 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-default-user" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.516613 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-server-dockercfg-hsns6" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.516969 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-notifications-svc" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.526756 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.638029 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.638137 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.638287 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.638345 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.638462 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.638491 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.638553 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.638622 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.638717 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.638752 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.638805 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk7vq\" (UniqueName: \"kubernetes.io/projected/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-kube-api-access-tk7vq\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.740503 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.740546 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.740569 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.740585 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.740617 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.740639 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.740660 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.741207 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.741372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.741434 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.741463 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.741489 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk7vq\" (UniqueName: \"kubernetes.io/projected/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-kube-api-access-tk7vq\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.743105 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.743918 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.744162 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.744536 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.749126 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.751391 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.751880 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.752092 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.765221 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk7vq\" (UniqueName: \"kubernetes.io/projected/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-kube-api-access-tk7vq\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.766940 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fcf1eb6a-de6d-48ac-bbf5-4e015c452e18-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.773855 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18\") " pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.831704 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:19:37 crc kubenswrapper[4810]: I0930 08:19:37.986888 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 08:19:37 crc kubenswrapper[4810]: W0930 08:19:37.998414 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf638129_9405_48c6_8458_143b297363fd.slice/crio-137f0c7fece1e342a6e0a26f2c3598092749e6b180d83bec687bf69776fa0d5e WatchSource:0}: Error finding container 137f0c7fece1e342a6e0a26f2c3598092749e6b180d83bec687bf69776fa0d5e: Status 404 returned error can't find the container with id 137f0c7fece1e342a6e0a26f2c3598092749e6b180d83bec687bf69776fa0d5e Sep 30 08:19:38 crc kubenswrapper[4810]: I0930 08:19:38.183607 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5bb3b225-5221-417f-9914-737723ac52aa","Type":"ContainerStarted","Data":"c3c7f0ab98f0a063109b6ecd0d4261b59e1118b8cc965b3ff513d4873de959b2"} Sep 30 08:19:38 crc kubenswrapper[4810]: I0930 08:19:38.184805 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cf638129-9405-48c6-8458-143b297363fd","Type":"ContainerStarted","Data":"137f0c7fece1e342a6e0a26f2c3598092749e6b180d83bec687bf69776fa0d5e"} Sep 30 08:19:38 crc kubenswrapper[4810]: I0930 08:19:38.186667 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" event={"ID":"7601717f-f941-49aa-bb26-309cfeb9401d","Type":"ContainerStarted","Data":"8d8758a484d8872511786184de235b5de21f1b47ca1977cf385e3e7367cdc501"} Sep 30 08:19:38 crc kubenswrapper[4810]: I0930 08:19:38.408719 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.431637 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.438206 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.440770 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-k6mfq" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.441244 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.441373 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.441434 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.442382 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.451829 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.458146 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.491291 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.491343 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.491376 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-kolla-config\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.491409 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzvzk\" (UniqueName: \"kubernetes.io/projected/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-kube-api-access-jzvzk\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.491431 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-secrets\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.491459 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.491477 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.491509 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.491530 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-config-data-default\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.593150 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.593226 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-kolla-config\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.593285 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzvzk\" (UniqueName: \"kubernetes.io/projected/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-kube-api-access-jzvzk\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.593313 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-secrets\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.593355 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.593384 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.593428 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.593455 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-config-data-default\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.593673 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.594558 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.596851 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-kolla-config\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.597488 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.597772 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.598411 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-config-data-default\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.602163 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.607020 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-secrets\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.612287 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.624750 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzvzk\" (UniqueName: \"kubernetes.io/projected/3ca490c3-ff2f-4553-9e66-4fb456d3a1f2-kube-api-access-jzvzk\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.628384 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2\") " pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.761011 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.806430 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.809042 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.816622 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.816994 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-p85v5" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.817562 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.817917 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.840295 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.899991 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cb4aa97-f850-4e7f-a1e1-046bd5235109-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.900037 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cb4aa97-f850-4e7f-a1e1-046bd5235109-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.900104 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bp8g\" (UniqueName: \"kubernetes.io/projected/7cb4aa97-f850-4e7f-a1e1-046bd5235109-kube-api-access-8bp8g\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.900151 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7cb4aa97-f850-4e7f-a1e1-046bd5235109-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.900244 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cb4aa97-f850-4e7f-a1e1-046bd5235109-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.900321 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.900345 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cb4aa97-f850-4e7f-a1e1-046bd5235109-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.901097 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb4aa97-f850-4e7f-a1e1-046bd5235109-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:40 crc kubenswrapper[4810]: I0930 08:19:40.901333 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cb4aa97-f850-4e7f-a1e1-046bd5235109-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.003287 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cb4aa97-f850-4e7f-a1e1-046bd5235109-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.003365 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bp8g\" (UniqueName: \"kubernetes.io/projected/7cb4aa97-f850-4e7f-a1e1-046bd5235109-kube-api-access-8bp8g\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.003394 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7cb4aa97-f850-4e7f-a1e1-046bd5235109-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.003423 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cb4aa97-f850-4e7f-a1e1-046bd5235109-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.003444 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.003476 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cb4aa97-f850-4e7f-a1e1-046bd5235109-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.003527 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb4aa97-f850-4e7f-a1e1-046bd5235109-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.003592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cb4aa97-f850-4e7f-a1e1-046bd5235109-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.003624 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cb4aa97-f850-4e7f-a1e1-046bd5235109-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.003760 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cb4aa97-f850-4e7f-a1e1-046bd5235109-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.003897 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.004461 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cb4aa97-f850-4e7f-a1e1-046bd5235109-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.008359 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cb4aa97-f850-4e7f-a1e1-046bd5235109-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.015898 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7cb4aa97-f850-4e7f-a1e1-046bd5235109-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.016851 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cb4aa97-f850-4e7f-a1e1-046bd5235109-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.025146 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb4aa97-f850-4e7f-a1e1-046bd5235109-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.026470 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cb4aa97-f850-4e7f-a1e1-046bd5235109-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.032771 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bp8g\" (UniqueName: \"kubernetes.io/projected/7cb4aa97-f850-4e7f-a1e1-046bd5235109-kube-api-access-8bp8g\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.036621 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cb4aa97-f850-4e7f-a1e1-046bd5235109\") " pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.143982 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.217749 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.219064 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.231011 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-26ntx" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.231181 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.231237 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.245763 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.314841 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f947e702-a475-413a-8e34-08702649bdae-config-data\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.317879 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk2v2\" (UniqueName: \"kubernetes.io/projected/f947e702-a475-413a-8e34-08702649bdae-kube-api-access-pk2v2\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.318157 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f947e702-a475-413a-8e34-08702649bdae-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.318246 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f947e702-a475-413a-8e34-08702649bdae-kolla-config\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.318376 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f947e702-a475-413a-8e34-08702649bdae-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.421212 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f947e702-a475-413a-8e34-08702649bdae-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.421256 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f947e702-a475-413a-8e34-08702649bdae-kolla-config\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.421292 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f947e702-a475-413a-8e34-08702649bdae-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.421310 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f947e702-a475-413a-8e34-08702649bdae-config-data\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.421397 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk2v2\" (UniqueName: \"kubernetes.io/projected/f947e702-a475-413a-8e34-08702649bdae-kube-api-access-pk2v2\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.422972 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f947e702-a475-413a-8e34-08702649bdae-kolla-config\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.423993 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f947e702-a475-413a-8e34-08702649bdae-config-data\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.431963 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f947e702-a475-413a-8e34-08702649bdae-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.436509 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f947e702-a475-413a-8e34-08702649bdae-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.438817 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk2v2\" (UniqueName: \"kubernetes.io/projected/f947e702-a475-413a-8e34-08702649bdae-kube-api-access-pk2v2\") pod \"memcached-0\" (UID: \"f947e702-a475-413a-8e34-08702649bdae\") " pod="openstack/memcached-0" Sep 30 08:19:41 crc kubenswrapper[4810]: I0930 08:19:41.564869 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 30 08:19:42 crc kubenswrapper[4810]: I0930 08:19:42.781328 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 08:19:42 crc kubenswrapper[4810]: I0930 08:19:42.782647 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 08:19:42 crc kubenswrapper[4810]: I0930 08:19:42.786567 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-7tnz7" Sep 30 08:19:42 crc kubenswrapper[4810]: I0930 08:19:42.797852 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 08:19:42 crc kubenswrapper[4810]: I0930 08:19:42.845134 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fspl9\" (UniqueName: \"kubernetes.io/projected/1218e0c8-f487-48d3-ae27-2e7b3aa4d07e-kube-api-access-fspl9\") pod \"kube-state-metrics-0\" (UID: \"1218e0c8-f487-48d3-ae27-2e7b3aa4d07e\") " pod="openstack/kube-state-metrics-0" Sep 30 08:19:42 crc kubenswrapper[4810]: I0930 08:19:42.946806 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fspl9\" (UniqueName: \"kubernetes.io/projected/1218e0c8-f487-48d3-ae27-2e7b3aa4d07e-kube-api-access-fspl9\") pod \"kube-state-metrics-0\" (UID: \"1218e0c8-f487-48d3-ae27-2e7b3aa4d07e\") " pod="openstack/kube-state-metrics-0" Sep 30 08:19:42 crc kubenswrapper[4810]: I0930 08:19:42.975950 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fspl9\" (UniqueName: \"kubernetes.io/projected/1218e0c8-f487-48d3-ae27-2e7b3aa4d07e-kube-api-access-fspl9\") pod \"kube-state-metrics-0\" (UID: \"1218e0c8-f487-48d3-ae27-2e7b3aa4d07e\") " pod="openstack/kube-state-metrics-0" Sep 30 08:19:43 crc kubenswrapper[4810]: I0930 08:19:43.108630 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.130139 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.138873 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.139073 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.143691 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.143856 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-t6zzk" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.143976 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.144077 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.144219 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.151817 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.270573 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-config\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.270624 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e9cd2658-4276-4957-9941-ab397d3e0d5f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.270794 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.270892 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.270967 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.271146 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.271186 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw8c9\" (UniqueName: \"kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-kube-api-access-vw8c9\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.271216 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e9cd2658-4276-4957-9941-ab397d3e0d5f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.373516 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.373606 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.373657 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.373755 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.373791 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw8c9\" (UniqueName: \"kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-kube-api-access-vw8c9\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.373815 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e9cd2658-4276-4957-9941-ab397d3e0d5f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.373884 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-config\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.373912 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e9cd2658-4276-4957-9941-ab397d3e0d5f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.384171 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e9cd2658-4276-4957-9941-ab397d3e0d5f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.391294 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.393677 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-config\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.395598 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.397019 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e9cd2658-4276-4957-9941-ab397d3e0d5f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.399346 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.399383 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b8ca484aa6b3c023d84489054d30b21e25eca5ebf5935b3c554e8125ba6153af/globalmount\"" pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.404299 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw8c9\" (UniqueName: \"kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-kube-api-access-vw8c9\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.410580 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.453895 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:44 crc kubenswrapper[4810]: I0930 08:19:44.475628 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 08:19:45 crc kubenswrapper[4810]: I0930 08:19:45.912042 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:19:45 crc kubenswrapper[4810]: I0930 08:19:45.912402 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.104585 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9q2d8"] Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.105971 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.109107 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.109735 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.109766 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-tgv2m" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.132653 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9q2d8"] Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.140000 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-mt698"] Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.142189 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.157011 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mt698"] Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235247 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/af34271b-1a44-4240-88ee-b86490744353-var-log-ovn\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235304 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxtvk\" (UniqueName: \"kubernetes.io/projected/af34271b-1a44-4240-88ee-b86490744353-kube-api-access-bxtvk\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235339 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/af34271b-1a44-4240-88ee-b86490744353-ovn-controller-tls-certs\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235424 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af34271b-1a44-4240-88ee-b86490744353-combined-ca-bundle\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235442 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-etc-ovs\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235510 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9scxb\" (UniqueName: \"kubernetes.io/projected/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-kube-api-access-9scxb\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235527 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-var-log\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235544 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af34271b-1a44-4240-88ee-b86490744353-scripts\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235608 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-var-lib\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235632 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-var-run\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235661 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/af34271b-1a44-4240-88ee-b86490744353-var-run-ovn\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235680 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/af34271b-1a44-4240-88ee-b86490744353-var-run\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.235699 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-scripts\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.336827 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-var-lib\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.336872 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-var-run\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.336911 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/af34271b-1a44-4240-88ee-b86490744353-var-run-ovn\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.336930 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/af34271b-1a44-4240-88ee-b86490744353-var-run\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.336949 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-scripts\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.336980 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/af34271b-1a44-4240-88ee-b86490744353-var-log-ovn\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.336999 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxtvk\" (UniqueName: \"kubernetes.io/projected/af34271b-1a44-4240-88ee-b86490744353-kube-api-access-bxtvk\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.337019 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/af34271b-1a44-4240-88ee-b86490744353-ovn-controller-tls-certs\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.337044 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af34271b-1a44-4240-88ee-b86490744353-combined-ca-bundle\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.337060 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-etc-ovs\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.337124 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9scxb\" (UniqueName: \"kubernetes.io/projected/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-kube-api-access-9scxb\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.337159 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-var-log\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.337191 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af34271b-1a44-4240-88ee-b86490744353-scripts\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.345139 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/af34271b-1a44-4240-88ee-b86490744353-var-log-ovn\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.345227 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/af34271b-1a44-4240-88ee-b86490744353-var-run\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.351818 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-etc-ovs\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.352057 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af34271b-1a44-4240-88ee-b86490744353-combined-ca-bundle\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.352451 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/af34271b-1a44-4240-88ee-b86490744353-ovn-controller-tls-certs\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.353188 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af34271b-1a44-4240-88ee-b86490744353-scripts\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.353942 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-scripts\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.356537 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-var-log\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.356571 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/af34271b-1a44-4240-88ee-b86490744353-var-run-ovn\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.356710 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-var-lib\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.356826 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-var-run\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.362889 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9scxb\" (UniqueName: \"kubernetes.io/projected/384b3d69-0ca1-4729-86f7-6d4455fb7ccc-kube-api-access-9scxb\") pod \"ovn-controller-ovs-mt698\" (UID: \"384b3d69-0ca1-4729-86f7-6d4455fb7ccc\") " pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.363728 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxtvk\" (UniqueName: \"kubernetes.io/projected/af34271b-1a44-4240-88ee-b86490744353-kube-api-access-bxtvk\") pod \"ovn-controller-9q2d8\" (UID: \"af34271b-1a44-4240-88ee-b86490744353\") " pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.423897 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8" Sep 30 08:19:47 crc kubenswrapper[4810]: I0930 08:19:47.460636 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.795955 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.797371 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.800628 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.800716 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.800642 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.800904 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-ln7sb" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.802244 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.816632 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.875388 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c94c8fec-2762-469d-86f9-a4aca5c242e5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.875436 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.875475 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c94c8fec-2762-469d-86f9-a4aca5c242e5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.875501 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blk7t\" (UniqueName: \"kubernetes.io/projected/c94c8fec-2762-469d-86f9-a4aca5c242e5-kube-api-access-blk7t\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.875629 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c94c8fec-2762-469d-86f9-a4aca5c242e5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.875738 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c94c8fec-2762-469d-86f9-a4aca5c242e5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.875766 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c94c8fec-2762-469d-86f9-a4aca5c242e5-config\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.875828 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c94c8fec-2762-469d-86f9-a4aca5c242e5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.977442 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c94c8fec-2762-469d-86f9-a4aca5c242e5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.977496 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.977548 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c94c8fec-2762-469d-86f9-a4aca5c242e5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.977579 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blk7t\" (UniqueName: \"kubernetes.io/projected/c94c8fec-2762-469d-86f9-a4aca5c242e5-kube-api-access-blk7t\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.977624 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c94c8fec-2762-469d-86f9-a4aca5c242e5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.977668 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c94c8fec-2762-469d-86f9-a4aca5c242e5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.977695 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c94c8fec-2762-469d-86f9-a4aca5c242e5-config\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.977723 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c94c8fec-2762-469d-86f9-a4aca5c242e5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.977793 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.978234 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c94c8fec-2762-469d-86f9-a4aca5c242e5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.978957 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c94c8fec-2762-469d-86f9-a4aca5c242e5-config\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.979053 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c94c8fec-2762-469d-86f9-a4aca5c242e5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.984031 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c94c8fec-2762-469d-86f9-a4aca5c242e5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.984572 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c94c8fec-2762-469d-86f9-a4aca5c242e5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.985262 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c94c8fec-2762-469d-86f9-a4aca5c242e5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:49 crc kubenswrapper[4810]: I0930 08:19:49.996039 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blk7t\" (UniqueName: \"kubernetes.io/projected/c94c8fec-2762-469d-86f9-a4aca5c242e5-kube-api-access-blk7t\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.004262 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c94c8fec-2762-469d-86f9-a4aca5c242e5\") " pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.119720 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.719920 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.724693 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.729515 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.729769 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.731814 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-jmp6z" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.732110 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.738099 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.791518 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.791853 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4jp4\" (UniqueName: \"kubernetes.io/projected/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-kube-api-access-d4jp4\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.791975 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.792046 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.792151 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-config\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.792258 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.792389 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.792491 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.894735 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.894790 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.894902 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-config\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.895025 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.895045 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.895073 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.895126 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.895208 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4jp4\" (UniqueName: \"kubernetes.io/projected/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-kube-api-access-d4jp4\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.895948 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.896138 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.897638 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.898064 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-config\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.900675 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.901816 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.902986 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.918204 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4jp4\" (UniqueName: \"kubernetes.io/projected/ba947eb7-f01f-40ef-b276-b5b73fee9e0a-kube-api-access-d4jp4\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:50 crc kubenswrapper[4810]: I0930 08:19:50.921233 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ba947eb7-f01f-40ef-b276-b5b73fee9e0a\") " pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:51 crc kubenswrapper[4810]: I0930 08:19:51.064244 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 30 08:19:51 crc kubenswrapper[4810]: W0930 08:19:51.946579 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcf1eb6a_de6d_48ac_bbf5_4e015c452e18.slice/crio-7b53188c33e0bfeb10284b37e32f6ee20e8454dd6026d2a5c8a85235dfd67e1a WatchSource:0}: Error finding container 7b53188c33e0bfeb10284b37e32f6ee20e8454dd6026d2a5c8a85235dfd67e1a: Status 404 returned error can't find the container with id 7b53188c33e0bfeb10284b37e32f6ee20e8454dd6026d2a5c8a85235dfd67e1a Sep 30 08:19:52 crc kubenswrapper[4810]: I0930 08:19:52.359247 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18","Type":"ContainerStarted","Data":"7b53188c33e0bfeb10284b37e32f6ee20e8454dd6026d2a5c8a85235dfd67e1a"} Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.372653 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.373105 4810 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.373307 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.41:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xww7s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5cccfc6d89-q2n82_openstack(b3ed1d67-4c86-49f1-8081-b89bc5d210d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.374458 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" podUID="b3ed1d67-4c86-49f1-8081-b89bc5d210d3" Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.466411 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.466478 4810 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.466578 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.41:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h5lqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5547dd57d7-76dtr_openstack(b5d305fa-0c3f-48b1-b17b-4f543398d53d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.467750 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" podUID="b5d305fa-0c3f-48b1-b17b-4f543398d53d" Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.822910 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.823242 4810 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.823378 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.41:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5bffj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5dc5bf6869-qgznr_openstack(fd327763-4320-40ff-8d03-7f48ffbfdc1a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 08:19:59 crc kubenswrapper[4810]: E0930 08:19:59.824546 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" podUID="fd327763-4320-40ff-8d03-7f48ffbfdc1a" Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.855177 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.893062 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.930316 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.976090 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5lqz\" (UniqueName: \"kubernetes.io/projected/b5d305fa-0c3f-48b1-b17b-4f543398d53d-kube-api-access-h5lqz\") pod \"b5d305fa-0c3f-48b1-b17b-4f543398d53d\" (UID: \"b5d305fa-0c3f-48b1-b17b-4f543398d53d\") " Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.976146 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-dns-svc\") pod \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.976211 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d305fa-0c3f-48b1-b17b-4f543398d53d-config\") pod \"b5d305fa-0c3f-48b1-b17b-4f543398d53d\" (UID: \"b5d305fa-0c3f-48b1-b17b-4f543398d53d\") " Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.976240 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bffj\" (UniqueName: \"kubernetes.io/projected/fd327763-4320-40ff-8d03-7f48ffbfdc1a-kube-api-access-5bffj\") pod \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.976301 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-config\") pod \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.976342 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-dns-svc\") pod \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\" (UID: \"fd327763-4320-40ff-8d03-7f48ffbfdc1a\") " Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.976369 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-config\") pod \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.976391 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xww7s\" (UniqueName: \"kubernetes.io/projected/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-kube-api-access-xww7s\") pod \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\" (UID: \"b3ed1d67-4c86-49f1-8081-b89bc5d210d3\") " Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.978858 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fd327763-4320-40ff-8d03-7f48ffbfdc1a" (UID: "fd327763-4320-40ff-8d03-7f48ffbfdc1a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.979024 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-config" (OuterVolumeSpecName: "config") pod "b3ed1d67-4c86-49f1-8081-b89bc5d210d3" (UID: "b3ed1d67-4c86-49f1-8081-b89bc5d210d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.979261 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-config" (OuterVolumeSpecName: "config") pod "fd327763-4320-40ff-8d03-7f48ffbfdc1a" (UID: "fd327763-4320-40ff-8d03-7f48ffbfdc1a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.979589 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5d305fa-0c3f-48b1-b17b-4f543398d53d-config" (OuterVolumeSpecName: "config") pod "b5d305fa-0c3f-48b1-b17b-4f543398d53d" (UID: "b5d305fa-0c3f-48b1-b17b-4f543398d53d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.979617 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b3ed1d67-4c86-49f1-8081-b89bc5d210d3" (UID: "b3ed1d67-4c86-49f1-8081-b89bc5d210d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.981500 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd327763-4320-40ff-8d03-7f48ffbfdc1a-kube-api-access-5bffj" (OuterVolumeSpecName: "kube-api-access-5bffj") pod "fd327763-4320-40ff-8d03-7f48ffbfdc1a" (UID: "fd327763-4320-40ff-8d03-7f48ffbfdc1a"). InnerVolumeSpecName "kube-api-access-5bffj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.983330 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d305fa-0c3f-48b1-b17b-4f543398d53d-kube-api-access-h5lqz" (OuterVolumeSpecName: "kube-api-access-h5lqz") pod "b5d305fa-0c3f-48b1-b17b-4f543398d53d" (UID: "b5d305fa-0c3f-48b1-b17b-4f543398d53d"). InnerVolumeSpecName "kube-api-access-h5lqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:20:00 crc kubenswrapper[4810]: I0930 08:20:00.984477 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-kube-api-access-xww7s" (OuterVolumeSpecName: "kube-api-access-xww7s") pod "b3ed1d67-4c86-49f1-8081-b89bc5d210d3" (UID: "b3ed1d67-4c86-49f1-8081-b89bc5d210d3"). InnerVolumeSpecName "kube-api-access-xww7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.080079 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bffj\" (UniqueName: \"kubernetes.io/projected/fd327763-4320-40ff-8d03-7f48ffbfdc1a-kube-api-access-5bffj\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.080116 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.080125 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd327763-4320-40ff-8d03-7f48ffbfdc1a-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.080135 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.080143 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xww7s\" (UniqueName: \"kubernetes.io/projected/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-kube-api-access-xww7s\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.080151 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5lqz\" (UniqueName: \"kubernetes.io/projected/b5d305fa-0c3f-48b1-b17b-4f543398d53d-kube-api-access-h5lqz\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.080161 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ed1d67-4c86-49f1-8081-b89bc5d210d3-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.080172 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d305fa-0c3f-48b1-b17b-4f543398d53d-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.391245 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.398844 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9q2d8"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.420664 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.429163 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" event={"ID":"b5d305fa-0c3f-48b1-b17b-4f543398d53d","Type":"ContainerDied","Data":"b9e2c9e1dc3faad353fcdc7e88c85fea05807fcfbf1fdf63d8fc213f72a537dc"} Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.429260 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5547dd57d7-76dtr" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.431503 4810 generic.go:334] "Generic (PLEG): container finished" podID="9161c7fc-8a57-49cd-ae3d-7043933ecae6" containerID="16b9b2f935e21f4310d828cc7593052aeb74ebaa663ddcd5bbdbad81fa0bed71" exitCode=0 Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.431551 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" event={"ID":"9161c7fc-8a57-49cd-ae3d-7043933ecae6","Type":"ContainerDied","Data":"16b9b2f935e21f4310d828cc7593052aeb74ebaa663ddcd5bbdbad81fa0bed71"} Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.435005 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.435020 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dc5bf6869-qgznr" event={"ID":"fd327763-4320-40ff-8d03-7f48ffbfdc1a","Type":"ContainerDied","Data":"56af85a2724b5e7077952fdc7107b1cdf8472635bc1a1ba6d4a67dcfa5b4af1d"} Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.440701 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18","Type":"ContainerStarted","Data":"a551b215e477b57fc81ad143e736021bdd6bf8f56c9f5c9f1a0e3e22332ede28"} Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.443691 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5bb3b225-5221-417f-9914-737723ac52aa","Type":"ContainerStarted","Data":"dba7b116bca716cd0a076f7032ee15fd1e500f765b290f96f285fbc493d3400f"} Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.449215 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.449890 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cccfc6d89-q2n82" event={"ID":"b3ed1d67-4c86-49f1-8081-b89bc5d210d3","Type":"ContainerDied","Data":"cebd1e83ac902bf75074f42db393fffe1274286690f1428739f176493d10a432"} Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.470303 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cf638129-9405-48c6-8458-143b297363fd","Type":"ContainerStarted","Data":"8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21"} Sep 30 08:20:01 crc kubenswrapper[4810]: W0930 08:20:01.472924 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf34271b_1a44_4240_88ee_b86490744353.slice/crio-90891474f75e3fc2f996f00305ce591660152c9c77ffdb1798b518b7602b7905 WatchSource:0}: Error finding container 90891474f75e3fc2f996f00305ce591660152c9c77ffdb1798b518b7602b7905: Status 404 returned error can't find the container with id 90891474f75e3fc2f996f00305ce591660152c9c77ffdb1798b518b7602b7905 Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.473127 4810 generic.go:334] "Generic (PLEG): container finished" podID="7601717f-f941-49aa-bb26-309cfeb9401d" containerID="bc0f55a44eb1d2a6c99a1c6ae2c568445b637f08d7adbb0db8fee6ff072a319a" exitCode=0 Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.473159 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" event={"ID":"7601717f-f941-49aa-bb26-309cfeb9401d","Type":"ContainerDied","Data":"bc0f55a44eb1d2a6c99a1c6ae2c568445b637f08d7adbb0db8fee6ff072a319a"} Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.570760 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.603615 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5547dd57d7-76dtr"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.610130 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5547dd57d7-76dtr"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.639497 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.665075 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.727876 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.742724 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cccfc6d89-q2n82"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.750141 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cccfc6d89-q2n82"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.773341 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dc5bf6869-qgznr"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.782183 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dc5bf6869-qgznr"] Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.790802 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mt698"] Sep 30 08:20:01 crc kubenswrapper[4810]: E0930 08:20:01.826411 4810 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Sep 30 08:20:01 crc kubenswrapper[4810]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/9161c7fc-8a57-49cd-ae3d-7043933ecae6/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Sep 30 08:20:01 crc kubenswrapper[4810]: > podSandboxID="61440a7aac7f390d75ca97d3995c538037dc758e60136ab5244f1d62752de9cb" Sep 30 08:20:01 crc kubenswrapper[4810]: E0930 08:20:01.826608 4810 kuberuntime_manager.go:1274] "Unhandled Error" err=< Sep 30 08:20:01 crc kubenswrapper[4810]: container &Container{Name:dnsmasq-dns,Image:38.102.83.41:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l6wt5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-75f7846495-qsb5s_openstack(9161c7fc-8a57-49cd-ae3d-7043933ecae6): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/9161c7fc-8a57-49cd-ae3d-7043933ecae6/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Sep 30 08:20:01 crc kubenswrapper[4810]: > logger="UnhandledError" Sep 30 08:20:01 crc kubenswrapper[4810]: E0930 08:20:01.827824 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/9161c7fc-8a57-49cd-ae3d-7043933ecae6/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" podUID="9161c7fc-8a57-49cd-ae3d-7043933ecae6" Sep 30 08:20:01 crc kubenswrapper[4810]: I0930 08:20:01.891617 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 08:20:01 crc kubenswrapper[4810]: W0930 08:20:01.900976 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba947eb7_f01f_40ef_b276_b5b73fee9e0a.slice/crio-eb977e019ea4e58185146554346cda7c3537d8cb4b6cc6755903d997e5ffd7fd WatchSource:0}: Error finding container eb977e019ea4e58185146554346cda7c3537d8cb4b6cc6755903d997e5ffd7fd: Status 404 returned error can't find the container with id eb977e019ea4e58185146554346cda7c3537d8cb4b6cc6755903d997e5ffd7fd Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.487871 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ba947eb7-f01f-40ef-b276-b5b73fee9e0a","Type":"ContainerStarted","Data":"eb977e019ea4e58185146554346cda7c3537d8cb4b6cc6755903d997e5ffd7fd"} Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.489130 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f947e702-a475-413a-8e34-08702649bdae","Type":"ContainerStarted","Data":"ec2f4db59e21c118214699675ebcd6fcd221df809eb958934e6709984d25086b"} Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.492492 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" event={"ID":"7601717f-f941-49aa-bb26-309cfeb9401d","Type":"ContainerStarted","Data":"586940b180c05e469c9e8b328fd16fe6ac2373eb3683b3bbe75d47cf9024eacc"} Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.493410 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.494855 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8" event={"ID":"af34271b-1a44-4240-88ee-b86490744353","Type":"ContainerStarted","Data":"90891474f75e3fc2f996f00305ce591660152c9c77ffdb1798b518b7602b7905"} Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.496134 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cb4aa97-f850-4e7f-a1e1-046bd5235109","Type":"ContainerStarted","Data":"0d69a25a21e3d55c9af5690bfd26ab6bfae24ad3a37c2e74907f96865cb5c78d"} Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.500847 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1218e0c8-f487-48d3-ae27-2e7b3aa4d07e","Type":"ContainerStarted","Data":"584a5c9639e46b31a8b121672e99c6a46c246d0a081979d0317e33b3e042b9a7"} Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.501887 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e9cd2658-4276-4957-9941-ab397d3e0d5f","Type":"ContainerStarted","Data":"fd53bdebccb0da1f617276a46b96b776d9408f63afda4bbf1230aaa839f976d0"} Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.503306 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mt698" event={"ID":"384b3d69-0ca1-4729-86f7-6d4455fb7ccc","Type":"ContainerStarted","Data":"5df64ed9bfda3e4dee19c327b5a056cecbdb555ce13c9eb113b2f8c353356e42"} Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.504914 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c94c8fec-2762-469d-86f9-a4aca5c242e5","Type":"ContainerStarted","Data":"a2cb05dd162b9706841138e13d2426943cc09afc9479f983081534ca74bcbae2"} Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.512156 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" podStartSLOduration=2.8979565149999997 podStartE2EDuration="26.512142248s" podCreationTimestamp="2025-09-30 08:19:36 +0000 UTC" firstStartedPulling="2025-09-30 08:19:37.240644429 +0000 UTC m=+1000.692843696" lastFinishedPulling="2025-09-30 08:20:00.854830162 +0000 UTC m=+1024.307029429" observedRunningTime="2025-09-30 08:20:02.5074245 +0000 UTC m=+1025.959623767" watchObservedRunningTime="2025-09-30 08:20:02.512142248 +0000 UTC m=+1025.964341515" Sep 30 08:20:02 crc kubenswrapper[4810]: I0930 08:20:02.515775 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2","Type":"ContainerStarted","Data":"b5266a5f745be06243ee663020c2df667671e9c7a031353ef6885df337c7ad7e"} Sep 30 08:20:03 crc kubenswrapper[4810]: I0930 08:20:03.318169 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3ed1d67-4c86-49f1-8081-b89bc5d210d3" path="/var/lib/kubelet/pods/b3ed1d67-4c86-49f1-8081-b89bc5d210d3/volumes" Sep 30 08:20:03 crc kubenswrapper[4810]: I0930 08:20:03.318924 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5d305fa-0c3f-48b1-b17b-4f543398d53d" path="/var/lib/kubelet/pods/b5d305fa-0c3f-48b1-b17b-4f543398d53d/volumes" Sep 30 08:20:03 crc kubenswrapper[4810]: I0930 08:20:03.319373 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd327763-4320-40ff-8d03-7f48ffbfdc1a" path="/var/lib/kubelet/pods/fd327763-4320-40ff-8d03-7f48ffbfdc1a/volumes" Sep 30 08:20:06 crc kubenswrapper[4810]: I0930 08:20:06.755197 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:20:06 crc kubenswrapper[4810]: I0930 08:20:06.830157 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75f7846495-qsb5s"] Sep 30 08:20:09 crc kubenswrapper[4810]: I0930 08:20:09.582311 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" event={"ID":"9161c7fc-8a57-49cd-ae3d-7043933ecae6","Type":"ContainerStarted","Data":"1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994"} Sep 30 08:20:09 crc kubenswrapper[4810]: I0930 08:20:09.583336 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" podUID="9161c7fc-8a57-49cd-ae3d-7043933ecae6" containerName="dnsmasq-dns" containerID="cri-o://1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994" gracePeriod=10 Sep 30 08:20:09 crc kubenswrapper[4810]: I0930 08:20:09.583931 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:20:09 crc kubenswrapper[4810]: I0930 08:20:09.613850 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" podStartSLOduration=10.800503671 podStartE2EDuration="34.61383162s" podCreationTimestamp="2025-09-30 08:19:35 +0000 UTC" firstStartedPulling="2025-09-30 08:19:37.015712213 +0000 UTC m=+1000.467911480" lastFinishedPulling="2025-09-30 08:20:00.829040162 +0000 UTC m=+1024.281239429" observedRunningTime="2025-09-30 08:20:09.61003833 +0000 UTC m=+1033.062237607" watchObservedRunningTime="2025-09-30 08:20:09.61383162 +0000 UTC m=+1033.066030887" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.497774 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.581757 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-dns-svc\") pod \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.581812 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6wt5\" (UniqueName: \"kubernetes.io/projected/9161c7fc-8a57-49cd-ae3d-7043933ecae6-kube-api-access-l6wt5\") pod \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.581893 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-config\") pod \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\" (UID: \"9161c7fc-8a57-49cd-ae3d-7043933ecae6\") " Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.588448 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9161c7fc-8a57-49cd-ae3d-7043933ecae6-kube-api-access-l6wt5" (OuterVolumeSpecName: "kube-api-access-l6wt5") pod "9161c7fc-8a57-49cd-ae3d-7043933ecae6" (UID: "9161c7fc-8a57-49cd-ae3d-7043933ecae6"). InnerVolumeSpecName "kube-api-access-l6wt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.604941 4810 generic.go:334] "Generic (PLEG): container finished" podID="9161c7fc-8a57-49cd-ae3d-7043933ecae6" containerID="1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994" exitCode=0 Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.604989 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" event={"ID":"9161c7fc-8a57-49cd-ae3d-7043933ecae6","Type":"ContainerDied","Data":"1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994"} Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.604997 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.605019 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75f7846495-qsb5s" event={"ID":"9161c7fc-8a57-49cd-ae3d-7043933ecae6","Type":"ContainerDied","Data":"61440a7aac7f390d75ca97d3995c538037dc758e60136ab5244f1d62752de9cb"} Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.605043 4810 scope.go:117] "RemoveContainer" containerID="1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.635976 4810 scope.go:117] "RemoveContainer" containerID="16b9b2f935e21f4310d828cc7593052aeb74ebaa663ddcd5bbdbad81fa0bed71" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.683866 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6wt5\" (UniqueName: \"kubernetes.io/projected/9161c7fc-8a57-49cd-ae3d-7043933ecae6-kube-api-access-l6wt5\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.690938 4810 scope.go:117] "RemoveContainer" containerID="1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994" Sep 30 08:20:10 crc kubenswrapper[4810]: E0930 08:20:10.691349 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994\": container with ID starting with 1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994 not found: ID does not exist" containerID="1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.691393 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994"} err="failed to get container status \"1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994\": rpc error: code = NotFound desc = could not find container \"1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994\": container with ID starting with 1f75b04ad0f075fb479d5c8dfc1cd3de1527c339959df343af9d7688f20cf994 not found: ID does not exist" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.691418 4810 scope.go:117] "RemoveContainer" containerID="16b9b2f935e21f4310d828cc7593052aeb74ebaa663ddcd5bbdbad81fa0bed71" Sep 30 08:20:10 crc kubenswrapper[4810]: E0930 08:20:10.691803 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16b9b2f935e21f4310d828cc7593052aeb74ebaa663ddcd5bbdbad81fa0bed71\": container with ID starting with 16b9b2f935e21f4310d828cc7593052aeb74ebaa663ddcd5bbdbad81fa0bed71 not found: ID does not exist" containerID="16b9b2f935e21f4310d828cc7593052aeb74ebaa663ddcd5bbdbad81fa0bed71" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.691830 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16b9b2f935e21f4310d828cc7593052aeb74ebaa663ddcd5bbdbad81fa0bed71"} err="failed to get container status \"16b9b2f935e21f4310d828cc7593052aeb74ebaa663ddcd5bbdbad81fa0bed71\": rpc error: code = NotFound desc = could not find container \"16b9b2f935e21f4310d828cc7593052aeb74ebaa663ddcd5bbdbad81fa0bed71\": container with ID starting with 16b9b2f935e21f4310d828cc7593052aeb74ebaa663ddcd5bbdbad81fa0bed71 not found: ID does not exist" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.701860 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9161c7fc-8a57-49cd-ae3d-7043933ecae6" (UID: "9161c7fc-8a57-49cd-ae3d-7043933ecae6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.723320 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-config" (OuterVolumeSpecName: "config") pod "9161c7fc-8a57-49cd-ae3d-7043933ecae6" (UID: "9161c7fc-8a57-49cd-ae3d-7043933ecae6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.785060 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.785125 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9161c7fc-8a57-49cd-ae3d-7043933ecae6-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.943395 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75f7846495-qsb5s"] Sep 30 08:20:10 crc kubenswrapper[4810]: I0930 08:20:10.950014 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75f7846495-qsb5s"] Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.319378 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9161c7fc-8a57-49cd-ae3d-7043933ecae6" path="/var/lib/kubelet/pods/9161c7fc-8a57-49cd-ae3d-7043933ecae6/volumes" Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.613517 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c94c8fec-2762-469d-86f9-a4aca5c242e5","Type":"ContainerStarted","Data":"35e0fcde6139f66bc663ab3f40fcb704c89fdd3328afe7a9607846d1aff4b844"} Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.614679 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cb4aa97-f850-4e7f-a1e1-046bd5235109","Type":"ContainerStarted","Data":"faf03dfdbd92a5ec947277ac47ffcc4ccdbec7f33da4744535bbf55ed77a1bfd"} Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.616559 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ba947eb7-f01f-40ef-b276-b5b73fee9e0a","Type":"ContainerStarted","Data":"98f35fc404a947c22100b0accdf7502be5abed892b92270c38bdbe59e6c10f3e"} Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.617936 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f947e702-a475-413a-8e34-08702649bdae","Type":"ContainerStarted","Data":"1d69f35143e34352a1547eca3a5a8a22a393cd2cf60311fb396e3d8756b7cf7d"} Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.618450 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.619675 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mt698" event={"ID":"384b3d69-0ca1-4729-86f7-6d4455fb7ccc","Type":"ContainerStarted","Data":"0f016c28059370a33f3c41686bf772fe5aaea6bd423aea4137a7147493b45dcf"} Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.622233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8" event={"ID":"af34271b-1a44-4240-88ee-b86490744353","Type":"ContainerStarted","Data":"6fdae3117f6815f55c41a38f8a76257f60cb29b491370f2e57374dd2aeb2f3e0"} Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.623540 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2","Type":"ContainerStarted","Data":"700445b48f9752d8bf4affa8a82fbcd516f7c05c7e3292183d4d078dfd37c443"} Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.627700 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1218e0c8-f487-48d3-ae27-2e7b3aa4d07e","Type":"ContainerStarted","Data":"dcb15d23ac7107388574027e20fb65adbf18bcb2018884e5a58b6b6967ddcfd1"} Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.628101 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.727924 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=24.198623527 podStartE2EDuration="30.727892098s" podCreationTimestamp="2025-09-30 08:19:41 +0000 UTC" firstStartedPulling="2025-09-30 08:20:01.746955554 +0000 UTC m=+1025.199154821" lastFinishedPulling="2025-09-30 08:20:08.276224085 +0000 UTC m=+1031.728423392" observedRunningTime="2025-09-30 08:20:11.726394515 +0000 UTC m=+1035.178593792" watchObservedRunningTime="2025-09-30 08:20:11.727892098 +0000 UTC m=+1035.180091375" Sep 30 08:20:11 crc kubenswrapper[4810]: I0930 08:20:11.756751 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=20.807269584 podStartE2EDuration="29.756725966s" podCreationTimestamp="2025-09-30 08:19:42 +0000 UTC" firstStartedPulling="2025-09-30 08:20:01.503015326 +0000 UTC m=+1024.955214593" lastFinishedPulling="2025-09-30 08:20:10.452471708 +0000 UTC m=+1033.904670975" observedRunningTime="2025-09-30 08:20:11.750916087 +0000 UTC m=+1035.203115354" watchObservedRunningTime="2025-09-30 08:20:11.756725966 +0000 UTC m=+1035.208925253" Sep 30 08:20:15 crc kubenswrapper[4810]: I0930 08:20:15.683254 4810 generic.go:334] "Generic (PLEG): container finished" podID="384b3d69-0ca1-4729-86f7-6d4455fb7ccc" containerID="0f016c28059370a33f3c41686bf772fe5aaea6bd423aea4137a7147493b45dcf" exitCode=0 Sep 30 08:20:15 crc kubenswrapper[4810]: I0930 08:20:15.683837 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mt698" event={"ID":"384b3d69-0ca1-4729-86f7-6d4455fb7ccc","Type":"ContainerDied","Data":"0f016c28059370a33f3c41686bf772fe5aaea6bd423aea4137a7147493b45dcf"} Sep 30 08:20:15 crc kubenswrapper[4810]: I0930 08:20:15.911588 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:20:15 crc kubenswrapper[4810]: I0930 08:20:15.911914 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:20:16 crc kubenswrapper[4810]: I0930 08:20:16.566284 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Sep 30 08:20:22 crc kubenswrapper[4810]: I0930 08:20:22.765515 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mt698" event={"ID":"384b3d69-0ca1-4729-86f7-6d4455fb7ccc","Type":"ContainerStarted","Data":"6e4ee0af0670190f6d72c33a890a526733b2136a8c97dd178ace52acd44886e4"} Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.154731 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.232536 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-777c66bf5c-8p229"] Sep 30 08:20:23 crc kubenswrapper[4810]: E0930 08:20:23.232895 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9161c7fc-8a57-49cd-ae3d-7043933ecae6" containerName="init" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.232905 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9161c7fc-8a57-49cd-ae3d-7043933ecae6" containerName="init" Sep 30 08:20:23 crc kubenswrapper[4810]: E0930 08:20:23.232933 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9161c7fc-8a57-49cd-ae3d-7043933ecae6" containerName="dnsmasq-dns" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.232939 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9161c7fc-8a57-49cd-ae3d-7043933ecae6" containerName="dnsmasq-dns" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.233094 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9161c7fc-8a57-49cd-ae3d-7043933ecae6" containerName="dnsmasq-dns" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.233974 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.258111 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-777c66bf5c-8p229"] Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.326229 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-498rl\" (UniqueName: \"kubernetes.io/projected/08ffde2b-6c3c-499f-b976-5aec9110596a-kube-api-access-498rl\") pod \"dnsmasq-dns-777c66bf5c-8p229\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.326379 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-dns-svc\") pod \"dnsmasq-dns-777c66bf5c-8p229\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.326520 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-config\") pod \"dnsmasq-dns-777c66bf5c-8p229\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.427689 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-config\") pod \"dnsmasq-dns-777c66bf5c-8p229\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.427978 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-498rl\" (UniqueName: \"kubernetes.io/projected/08ffde2b-6c3c-499f-b976-5aec9110596a-kube-api-access-498rl\") pod \"dnsmasq-dns-777c66bf5c-8p229\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.428042 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-dns-svc\") pod \"dnsmasq-dns-777c66bf5c-8p229\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.428914 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-config\") pod \"dnsmasq-dns-777c66bf5c-8p229\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.428980 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-dns-svc\") pod \"dnsmasq-dns-777c66bf5c-8p229\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.447838 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-498rl\" (UniqueName: \"kubernetes.io/projected/08ffde2b-6c3c-499f-b976-5aec9110596a-kube-api-access-498rl\") pod \"dnsmasq-dns-777c66bf5c-8p229\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:23 crc kubenswrapper[4810]: I0930 08:20:23.595801 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.300034 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.307566 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.315076 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.315311 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.315480 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.315643 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-xmn7c" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.325326 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.448840 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4pck\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-kube-api-access-n4pck\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.448964 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.448996 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/edd67220-e97a-465e-9671-2cfcf760475a-lock\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.449025 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/edd67220-e97a-465e-9671-2cfcf760475a-cache\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.449051 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.549998 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.550053 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/edd67220-e97a-465e-9671-2cfcf760475a-lock\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.550085 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/edd67220-e97a-465e-9671-2cfcf760475a-cache\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.550114 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.550163 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4pck\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-kube-api-access-n4pck\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.550735 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.553096 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/edd67220-e97a-465e-9671-2cfcf760475a-lock\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.553408 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/edd67220-e97a-465e-9671-2cfcf760475a-cache\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: E0930 08:20:24.553537 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 08:20:24 crc kubenswrapper[4810]: E0930 08:20:24.553554 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 08:20:24 crc kubenswrapper[4810]: E0930 08:20:24.553614 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift podName:edd67220-e97a-465e-9671-2cfcf760475a nodeName:}" failed. No retries permitted until 2025-09-30 08:20:25.053594052 +0000 UTC m=+1048.505793319 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift") pod "swift-storage-0" (UID: "edd67220-e97a-465e-9671-2cfcf760475a") : configmap "swift-ring-files" not found Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.556456 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-zr8dp"] Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.557809 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.560786 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.561145 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.561450 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.577490 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4pck\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-kube-api-access-n4pck\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.587541 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zr8dp"] Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.595206 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.651744 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-combined-ca-bundle\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.652108 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-swiftconf\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.652150 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-scripts\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.652170 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf76de28-c7b8-474f-93a7-5af536e817e8-etc-swift\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.652212 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99mc9\" (UniqueName: \"kubernetes.io/projected/cf76de28-c7b8-474f-93a7-5af536e817e8-kube-api-access-99mc9\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.652255 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-dispersionconf\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.652330 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-ring-data-devices\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.756028 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-scripts\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.756062 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf76de28-c7b8-474f-93a7-5af536e817e8-etc-swift\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.756093 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99mc9\" (UniqueName: \"kubernetes.io/projected/cf76de28-c7b8-474f-93a7-5af536e817e8-kube-api-access-99mc9\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.756124 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-dispersionconf\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.756153 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-ring-data-devices\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.757029 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf76de28-c7b8-474f-93a7-5af536e817e8-etc-swift\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.757327 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-combined-ca-bundle\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.757362 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-swiftconf\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.758422 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-scripts\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.762614 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-ring-data-devices\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.769570 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-dispersionconf\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.777995 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-combined-ca-bundle\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.778471 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-swiftconf\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.780741 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99mc9\" (UniqueName: \"kubernetes.io/projected/cf76de28-c7b8-474f-93a7-5af536e817e8-kube-api-access-99mc9\") pod \"swift-ring-rebalance-zr8dp\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.788551 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e9cd2658-4276-4957-9941-ab397d3e0d5f","Type":"ContainerStarted","Data":"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8"} Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.830462 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mt698" event={"ID":"384b3d69-0ca1-4729-86f7-6d4455fb7ccc","Type":"ContainerStarted","Data":"bd0d0c6f5280e150b21d032c8ae5b5587702703cd50f18226830a657a61b5ff8"} Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.830543 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-9q2d8" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.831229 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.831285 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.856375 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-9q2d8" podStartSLOduration=30.638074879 podStartE2EDuration="37.856347999s" podCreationTimestamp="2025-09-30 08:19:47 +0000 UTC" firstStartedPulling="2025-09-30 08:20:01.502958485 +0000 UTC m=+1024.955157752" lastFinishedPulling="2025-09-30 08:20:08.721231585 +0000 UTC m=+1032.173430872" observedRunningTime="2025-09-30 08:20:24.855235386 +0000 UTC m=+1048.307434653" watchObservedRunningTime="2025-09-30 08:20:24.856347999 +0000 UTC m=+1048.308547266" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.880159 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-mt698" podStartSLOduration=30.97128715 podStartE2EDuration="37.8801394s" podCreationTimestamp="2025-09-30 08:19:47 +0000 UTC" firstStartedPulling="2025-09-30 08:20:01.78568347 +0000 UTC m=+1025.237882737" lastFinishedPulling="2025-09-30 08:20:08.69453572 +0000 UTC m=+1032.146734987" observedRunningTime="2025-09-30 08:20:24.879476181 +0000 UTC m=+1048.331675448" watchObservedRunningTime="2025-09-30 08:20:24.8801394 +0000 UTC m=+1048.332338657" Sep 30 08:20:24 crc kubenswrapper[4810]: I0930 08:20:24.959793 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:25 crc kubenswrapper[4810]: I0930 08:20:25.062242 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:25 crc kubenswrapper[4810]: E0930 08:20:25.062567 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 08:20:25 crc kubenswrapper[4810]: E0930 08:20:25.062616 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 08:20:25 crc kubenswrapper[4810]: E0930 08:20:25.062720 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift podName:edd67220-e97a-465e-9671-2cfcf760475a nodeName:}" failed. No retries permitted until 2025-09-30 08:20:26.062680164 +0000 UTC m=+1049.514879441 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift") pod "swift-storage-0" (UID: "edd67220-e97a-465e-9671-2cfcf760475a") : configmap "swift-ring-files" not found Sep 30 08:20:26 crc kubenswrapper[4810]: I0930 08:20:26.086154 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:26 crc kubenswrapper[4810]: E0930 08:20:26.086644 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 08:20:26 crc kubenswrapper[4810]: E0930 08:20:26.086660 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 08:20:26 crc kubenswrapper[4810]: E0930 08:20:26.086707 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift podName:edd67220-e97a-465e-9671-2cfcf760475a nodeName:}" failed. No retries permitted until 2025-09-30 08:20:28.086691579 +0000 UTC m=+1051.538890846 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift") pod "swift-storage-0" (UID: "edd67220-e97a-465e-9671-2cfcf760475a") : configmap "swift-ring-files" not found Sep 30 08:20:26 crc kubenswrapper[4810]: I0930 08:20:26.290356 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-777c66bf5c-8p229"] Sep 30 08:20:26 crc kubenswrapper[4810]: I0930 08:20:26.352017 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zr8dp"] Sep 30 08:20:26 crc kubenswrapper[4810]: W0930 08:20:26.728802 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08ffde2b_6c3c_499f_b976_5aec9110596a.slice/crio-e7b56c2d200298285f38d222ec2981f2b8fc27e4857301dd0fa0c75611e7ecd6 WatchSource:0}: Error finding container e7b56c2d200298285f38d222ec2981f2b8fc27e4857301dd0fa0c75611e7ecd6: Status 404 returned error can't find the container with id e7b56c2d200298285f38d222ec2981f2b8fc27e4857301dd0fa0c75611e7ecd6 Sep 30 08:20:26 crc kubenswrapper[4810]: I0930 08:20:26.845697 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" event={"ID":"08ffde2b-6c3c-499f-b976-5aec9110596a","Type":"ContainerStarted","Data":"e7b56c2d200298285f38d222ec2981f2b8fc27e4857301dd0fa0c75611e7ecd6"} Sep 30 08:20:26 crc kubenswrapper[4810]: I0930 08:20:26.846723 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zr8dp" event={"ID":"cf76de28-c7b8-474f-93a7-5af536e817e8","Type":"ContainerStarted","Data":"03ee4defcb039cff19db4cf82a2c3165daa68373bd332d0eece308e9cb9fedb3"} Sep 30 08:20:27 crc kubenswrapper[4810]: I0930 08:20:27.862106 4810 generic.go:334] "Generic (PLEG): container finished" podID="08ffde2b-6c3c-499f-b976-5aec9110596a" containerID="61edeae012f0b6bb4f2f37586092df9f0e7e4b3d826d99f5d94812912fa92eee" exitCode=0 Sep 30 08:20:27 crc kubenswrapper[4810]: I0930 08:20:27.862146 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" event={"ID":"08ffde2b-6c3c-499f-b976-5aec9110596a","Type":"ContainerDied","Data":"61edeae012f0b6bb4f2f37586092df9f0e7e4b3d826d99f5d94812912fa92eee"} Sep 30 08:20:27 crc kubenswrapper[4810]: I0930 08:20:27.866727 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c94c8fec-2762-469d-86f9-a4aca5c242e5","Type":"ContainerStarted","Data":"df8fca26f76daae2a66618af4efb7e82b2c2b8cde0915dde4d85054e09972cd2"} Sep 30 08:20:27 crc kubenswrapper[4810]: I0930 08:20:27.871248 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ba947eb7-f01f-40ef-b276-b5b73fee9e0a","Type":"ContainerStarted","Data":"bad094a119aa064c71df51658d28a46af00bcb2be3ed94f5c64032df58a35b5d"} Sep 30 08:20:27 crc kubenswrapper[4810]: I0930 08:20:27.910483 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=14.430566128 podStartE2EDuration="39.910465452s" podCreationTimestamp="2025-09-30 08:19:48 +0000 UTC" firstStartedPulling="2025-09-30 08:20:01.643787947 +0000 UTC m=+1025.095987214" lastFinishedPulling="2025-09-30 08:20:27.123687261 +0000 UTC m=+1050.575886538" observedRunningTime="2025-09-30 08:20:27.905509328 +0000 UTC m=+1051.357708595" watchObservedRunningTime="2025-09-30 08:20:27.910465452 +0000 UTC m=+1051.362664709" Sep 30 08:20:27 crc kubenswrapper[4810]: I0930 08:20:27.936465 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=13.757885359 podStartE2EDuration="38.936445917s" podCreationTimestamp="2025-09-30 08:19:49 +0000 UTC" firstStartedPulling="2025-09-30 08:20:01.904718579 +0000 UTC m=+1025.356917846" lastFinishedPulling="2025-09-30 08:20:27.083279127 +0000 UTC m=+1050.535478404" observedRunningTime="2025-09-30 08:20:27.929915527 +0000 UTC m=+1051.382114794" watchObservedRunningTime="2025-09-30 08:20:27.936445917 +0000 UTC m=+1051.388645184" Sep 30 08:20:28 crc kubenswrapper[4810]: I0930 08:20:28.121930 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:28 crc kubenswrapper[4810]: E0930 08:20:28.122098 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 08:20:28 crc kubenswrapper[4810]: E0930 08:20:28.122371 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 08:20:28 crc kubenswrapper[4810]: E0930 08:20:28.122431 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift podName:edd67220-e97a-465e-9671-2cfcf760475a nodeName:}" failed. No retries permitted until 2025-09-30 08:20:32.12240882 +0000 UTC m=+1055.574608087 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift") pod "swift-storage-0" (UID: "edd67220-e97a-465e-9671-2cfcf760475a") : configmap "swift-ring-files" not found Sep 30 08:20:28 crc kubenswrapper[4810]: I0930 08:20:28.883063 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" event={"ID":"08ffde2b-6c3c-499f-b976-5aec9110596a","Type":"ContainerStarted","Data":"9591a44deb02fa623e0e7d09d6fcc2ad3759ec2fe622e8fe11f99cd930d403bf"} Sep 30 08:20:28 crc kubenswrapper[4810]: I0930 08:20:28.883564 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:28 crc kubenswrapper[4810]: I0930 08:20:28.905012 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" podStartSLOduration=5.90499449 podStartE2EDuration="5.90499449s" podCreationTimestamp="2025-09-30 08:20:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:20:28.901250001 +0000 UTC m=+1052.353449268" watchObservedRunningTime="2025-09-30 08:20:28.90499449 +0000 UTC m=+1052.357193757" Sep 30 08:20:29 crc kubenswrapper[4810]: I0930 08:20:29.121216 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Sep 30 08:20:29 crc kubenswrapper[4810]: I0930 08:20:29.163007 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Sep 30 08:20:29 crc kubenswrapper[4810]: I0930 08:20:29.891447 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Sep 30 08:20:29 crc kubenswrapper[4810]: I0930 08:20:29.939608 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.064406 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.102991 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.264779 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-777c66bf5c-8p229"] Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.289992 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79dbfbd97-tsh5g"] Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.291999 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.294869 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.314429 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dbfbd97-tsh5g"] Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.324148 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-6wrgl"] Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.325248 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.338035 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.365999 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-6wrgl"] Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.482863 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b9a3337-b2c2-49c8-9755-81660bb506af-config\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.482967 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9a3337-b2c2-49c8-9755-81660bb506af-combined-ca-bundle\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.483024 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b9a3337-b2c2-49c8-9755-81660bb506af-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.483063 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-dns-svc\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.483092 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4b9a3337-b2c2-49c8-9755-81660bb506af-ovn-rundir\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.483128 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhv7b\" (UniqueName: \"kubernetes.io/projected/4b9a3337-b2c2-49c8-9755-81660bb506af-kube-api-access-vhv7b\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.483186 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvtwb\" (UniqueName: \"kubernetes.io/projected/8c9dead4-5ad1-40c7-8f76-2ebed2449991-kube-api-access-zvtwb\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.483234 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4b9a3337-b2c2-49c8-9755-81660bb506af-ovs-rundir\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.483299 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-config\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.483324 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-ovsdbserver-nb\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.584828 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b9a3337-b2c2-49c8-9755-81660bb506af-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.584909 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-dns-svc\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.584946 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4b9a3337-b2c2-49c8-9755-81660bb506af-ovn-rundir\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.584972 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhv7b\" (UniqueName: \"kubernetes.io/projected/4b9a3337-b2c2-49c8-9755-81660bb506af-kube-api-access-vhv7b\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.585021 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvtwb\" (UniqueName: \"kubernetes.io/projected/8c9dead4-5ad1-40c7-8f76-2ebed2449991-kube-api-access-zvtwb\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.585060 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4b9a3337-b2c2-49c8-9755-81660bb506af-ovs-rundir\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.585107 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-config\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.585128 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-ovsdbserver-nb\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.585156 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b9a3337-b2c2-49c8-9755-81660bb506af-config\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.585228 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9a3337-b2c2-49c8-9755-81660bb506af-combined-ca-bundle\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.585432 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4b9a3337-b2c2-49c8-9755-81660bb506af-ovs-rundir\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.585490 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4b9a3337-b2c2-49c8-9755-81660bb506af-ovn-rundir\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.585972 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b9a3337-b2c2-49c8-9755-81660bb506af-config\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.586054 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-config\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.586169 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-ovsdbserver-nb\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.586279 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-dns-svc\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.599149 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9a3337-b2c2-49c8-9755-81660bb506af-combined-ca-bundle\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.606196 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b9a3337-b2c2-49c8-9755-81660bb506af-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.614570 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhv7b\" (UniqueName: \"kubernetes.io/projected/4b9a3337-b2c2-49c8-9755-81660bb506af-kube-api-access-vhv7b\") pod \"ovn-controller-metrics-6wrgl\" (UID: \"4b9a3337-b2c2-49c8-9755-81660bb506af\") " pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.624982 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvtwb\" (UniqueName: \"kubernetes.io/projected/8c9dead4-5ad1-40c7-8f76-2ebed2449991-kube-api-access-zvtwb\") pod \"dnsmasq-dns-79dbfbd97-tsh5g\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.642635 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6wrgl" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.705980 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79dbfbd97-tsh5g"] Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.706638 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.726717 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c8b4c69d9-t2bvs"] Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.729871 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.733584 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.758722 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c8b4c69d9-t2bvs"] Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.891177 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-dns-svc\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.891221 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.891499 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-config\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.891662 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dgtb\" (UniqueName: \"kubernetes.io/projected/73deb073-9ff6-4ef6-8e08-98ecc936dca1-kube-api-access-8dgtb\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.891814 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.909354 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" podUID="08ffde2b-6c3c-499f-b976-5aec9110596a" containerName="dnsmasq-dns" containerID="cri-o://9591a44deb02fa623e0e7d09d6fcc2ad3759ec2fe622e8fe11f99cd930d403bf" gracePeriod=10 Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.909382 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.966823 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.993900 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-config\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.993953 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dgtb\" (UniqueName: \"kubernetes.io/projected/73deb073-9ff6-4ef6-8e08-98ecc936dca1-kube-api-access-8dgtb\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.993986 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.994058 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-dns-svc\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.994072 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.994894 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.995387 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-config\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.995543 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-dns-svc\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:30 crc kubenswrapper[4810]: I0930 08:20:30.995715 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.020386 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dgtb\" (UniqueName: \"kubernetes.io/projected/73deb073-9ff6-4ef6-8e08-98ecc936dca1-kube-api-access-8dgtb\") pod \"dnsmasq-dns-7c8b4c69d9-t2bvs\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.058986 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.129677 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.136317 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.141969 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.142004 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.141969 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.142104 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-6sfmg" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.143177 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.306323 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/306e4bf0-015f-4c7a-ba49-efda86e72442-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.306577 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/306e4bf0-015f-4c7a-ba49-efda86e72442-config\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.306630 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/306e4bf0-015f-4c7a-ba49-efda86e72442-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.306660 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/306e4bf0-015f-4c7a-ba49-efda86e72442-scripts\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.306704 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/306e4bf0-015f-4c7a-ba49-efda86e72442-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.306743 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/306e4bf0-015f-4c7a-ba49-efda86e72442-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.306769 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txfln\" (UniqueName: \"kubernetes.io/projected/306e4bf0-015f-4c7a-ba49-efda86e72442-kube-api-access-txfln\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.408049 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/306e4bf0-015f-4c7a-ba49-efda86e72442-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.408118 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/306e4bf0-015f-4c7a-ba49-efda86e72442-scripts\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.408210 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/306e4bf0-015f-4c7a-ba49-efda86e72442-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.408288 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/306e4bf0-015f-4c7a-ba49-efda86e72442-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.408314 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txfln\" (UniqueName: \"kubernetes.io/projected/306e4bf0-015f-4c7a-ba49-efda86e72442-kube-api-access-txfln\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.408373 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/306e4bf0-015f-4c7a-ba49-efda86e72442-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.408402 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/306e4bf0-015f-4c7a-ba49-efda86e72442-config\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.408850 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/306e4bf0-015f-4c7a-ba49-efda86e72442-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.409320 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/306e4bf0-015f-4c7a-ba49-efda86e72442-config\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.409861 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/306e4bf0-015f-4c7a-ba49-efda86e72442-scripts\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.414212 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/306e4bf0-015f-4c7a-ba49-efda86e72442-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.415752 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/306e4bf0-015f-4c7a-ba49-efda86e72442-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.418920 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/306e4bf0-015f-4c7a-ba49-efda86e72442-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.426139 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txfln\" (UniqueName: \"kubernetes.io/projected/306e4bf0-015f-4c7a-ba49-efda86e72442-kube-api-access-txfln\") pod \"ovn-northd-0\" (UID: \"306e4bf0-015f-4c7a-ba49-efda86e72442\") " pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.462052 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.928840 4810 generic.go:334] "Generic (PLEG): container finished" podID="5bb3b225-5221-417f-9914-737723ac52aa" containerID="dba7b116bca716cd0a076f7032ee15fd1e500f765b290f96f285fbc493d3400f" exitCode=0 Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.928935 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5bb3b225-5221-417f-9914-737723ac52aa","Type":"ContainerDied","Data":"dba7b116bca716cd0a076f7032ee15fd1e500f765b290f96f285fbc493d3400f"} Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.932811 4810 generic.go:334] "Generic (PLEG): container finished" podID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerID="e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8" exitCode=0 Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.932888 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e9cd2658-4276-4957-9941-ab397d3e0d5f","Type":"ContainerDied","Data":"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8"} Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.935359 4810 generic.go:334] "Generic (PLEG): container finished" podID="cf638129-9405-48c6-8458-143b297363fd" containerID="8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21" exitCode=0 Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.935426 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cf638129-9405-48c6-8458-143b297363fd","Type":"ContainerDied","Data":"8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21"} Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.938222 4810 generic.go:334] "Generic (PLEG): container finished" podID="08ffde2b-6c3c-499f-b976-5aec9110596a" containerID="9591a44deb02fa623e0e7d09d6fcc2ad3759ec2fe622e8fe11f99cd930d403bf" exitCode=0 Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.938285 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" event={"ID":"08ffde2b-6c3c-499f-b976-5aec9110596a","Type":"ContainerDied","Data":"9591a44deb02fa623e0e7d09d6fcc2ad3759ec2fe622e8fe11f99cd930d403bf"} Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.940721 4810 generic.go:334] "Generic (PLEG): container finished" podID="fcf1eb6a-de6d-48ac-bbf5-4e015c452e18" containerID="a551b215e477b57fc81ad143e736021bdd6bf8f56c9f5c9f1a0e3e22332ede28" exitCode=0 Sep 30 08:20:31 crc kubenswrapper[4810]: I0930 08:20:31.940824 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18","Type":"ContainerDied","Data":"a551b215e477b57fc81ad143e736021bdd6bf8f56c9f5c9f1a0e3e22332ede28"} Sep 30 08:20:32 crc kubenswrapper[4810]: I0930 08:20:32.123457 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:32 crc kubenswrapper[4810]: E0930 08:20:32.124601 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 08:20:32 crc kubenswrapper[4810]: E0930 08:20:32.125420 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 08:20:32 crc kubenswrapper[4810]: E0930 08:20:32.125559 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift podName:edd67220-e97a-465e-9671-2cfcf760475a nodeName:}" failed. No retries permitted until 2025-09-30 08:20:40.125539589 +0000 UTC m=+1063.577738936 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift") pod "swift-storage-0" (UID: "edd67220-e97a-465e-9671-2cfcf760475a") : configmap "swift-ring-files" not found Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.558947 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 30 08:20:33 crc kubenswrapper[4810]: W0930 08:20:33.569336 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod306e4bf0_015f_4c7a_ba49_efda86e72442.slice/crio-250fb4616896aad17c29f6b59aac052e15a04061ff4dd7cef48c0383eb280be8 WatchSource:0}: Error finding container 250fb4616896aad17c29f6b59aac052e15a04061ff4dd7cef48c0383eb280be8: Status 404 returned error can't find the container with id 250fb4616896aad17c29f6b59aac052e15a04061ff4dd7cef48c0383eb280be8 Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.575503 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79dbfbd97-tsh5g"] Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.746631 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.823301 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-6wrgl"] Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.852372 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c8b4c69d9-t2bvs"] Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.880960 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-498rl\" (UniqueName: \"kubernetes.io/projected/08ffde2b-6c3c-499f-b976-5aec9110596a-kube-api-access-498rl\") pod \"08ffde2b-6c3c-499f-b976-5aec9110596a\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.881002 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-config\") pod \"08ffde2b-6c3c-499f-b976-5aec9110596a\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.881063 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-dns-svc\") pod \"08ffde2b-6c3c-499f-b976-5aec9110596a\" (UID: \"08ffde2b-6c3c-499f-b976-5aec9110596a\") " Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.885859 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08ffde2b-6c3c-499f-b976-5aec9110596a-kube-api-access-498rl" (OuterVolumeSpecName: "kube-api-access-498rl") pod "08ffde2b-6c3c-499f-b976-5aec9110596a" (UID: "08ffde2b-6c3c-499f-b976-5aec9110596a"). InnerVolumeSpecName "kube-api-access-498rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.978555 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-config" (OuterVolumeSpecName: "config") pod "08ffde2b-6c3c-499f-b976-5aec9110596a" (UID: "08ffde2b-6c3c-499f-b976-5aec9110596a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.988301 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-498rl\" (UniqueName: \"kubernetes.io/projected/08ffde2b-6c3c-499f-b976-5aec9110596a-kube-api-access-498rl\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.988347 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:33 crc kubenswrapper[4810]: I0930 08:20:33.993167 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08ffde2b-6c3c-499f-b976-5aec9110596a" (UID: "08ffde2b-6c3c-499f-b976-5aec9110596a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.017538 4810 generic.go:334] "Generic (PLEG): container finished" podID="8c9dead4-5ad1-40c7-8f76-2ebed2449991" containerID="8b9d2848bf9b9b3622109d67971a004666e07a2906702ee98e963fc3c31bc76f" exitCode=0 Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.017596 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" event={"ID":"8c9dead4-5ad1-40c7-8f76-2ebed2449991","Type":"ContainerDied","Data":"8b9d2848bf9b9b3622109d67971a004666e07a2906702ee98e963fc3c31bc76f"} Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.017625 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" event={"ID":"8c9dead4-5ad1-40c7-8f76-2ebed2449991","Type":"ContainerStarted","Data":"469bacefc378a02c0911eeee1e28332be374cce8ffedfb816b15f091fb0954ec"} Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.022229 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" event={"ID":"73deb073-9ff6-4ef6-8e08-98ecc936dca1","Type":"ContainerStarted","Data":"0ce0d454fe6cc9ba1fd245afcc8a68433f89a31289824e3a69f79e30039ac7e2"} Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.026192 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5bb3b225-5221-417f-9914-737723ac52aa","Type":"ContainerStarted","Data":"cc31d0996668af86e34cd3c8dc8059da7acee4bb9d1c4051edba62b215337f5b"} Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.026965 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.031533 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6wrgl" event={"ID":"4b9a3337-b2c2-49c8-9755-81660bb506af","Type":"ContainerStarted","Data":"4f39f801c4a4c4cd14b160529035ef7f3c481f7373d9aa06d9f1614a656b3696"} Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.034478 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" event={"ID":"08ffde2b-6c3c-499f-b976-5aec9110596a","Type":"ContainerDied","Data":"e7b56c2d200298285f38d222ec2981f2b8fc27e4857301dd0fa0c75611e7ecd6"} Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.034516 4810 scope.go:117] "RemoveContainer" containerID="9591a44deb02fa623e0e7d09d6fcc2ad3759ec2fe622e8fe11f99cd930d403bf" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.034655 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.059598 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zr8dp" event={"ID":"cf76de28-c7b8-474f-93a7-5af536e817e8","Type":"ContainerStarted","Data":"98977281bf04b6773e81a3868aa50e6c5604dcb10433b7c39cf6f1c9906e4585"} Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.060783 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"306e4bf0-015f-4c7a-ba49-efda86e72442","Type":"ContainerStarted","Data":"250fb4616896aad17c29f6b59aac052e15a04061ff4dd7cef48c0383eb280be8"} Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.066737 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"fcf1eb6a-de6d-48ac-bbf5-4e015c452e18","Type":"ContainerStarted","Data":"24ad5603e2ff109236744eb09639f35bb34d05f68a186aa01f066b58e3ffe3f8"} Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.066959 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.069435 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cf638129-9405-48c6-8458-143b297363fd","Type":"ContainerStarted","Data":"a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f"} Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.070009 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.085834 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=35.743584377 podStartE2EDuration="59.085789918s" podCreationTimestamp="2025-09-30 08:19:35 +0000 UTC" firstStartedPulling="2025-09-30 08:19:37.485353028 +0000 UTC m=+1000.937552285" lastFinishedPulling="2025-09-30 08:20:00.827558559 +0000 UTC m=+1024.279757826" observedRunningTime="2025-09-30 08:20:34.07657677 +0000 UTC m=+1057.528776037" watchObservedRunningTime="2025-09-30 08:20:34.085789918 +0000 UTC m=+1057.537989185" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.092415 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ffde2b-6c3c-499f-b976-5aec9110596a-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.132803 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-zr8dp" podStartSLOduration=3.593510943 podStartE2EDuration="10.132778763s" podCreationTimestamp="2025-09-30 08:20:24 +0000 UTC" firstStartedPulling="2025-09-30 08:20:26.861372509 +0000 UTC m=+1050.313571776" lastFinishedPulling="2025-09-30 08:20:33.400640329 +0000 UTC m=+1056.852839596" observedRunningTime="2025-09-30 08:20:34.127782578 +0000 UTC m=+1057.579981845" watchObservedRunningTime="2025-09-30 08:20:34.132778763 +0000 UTC m=+1057.584978030" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.144217 4810 scope.go:117] "RemoveContainer" containerID="61edeae012f0b6bb4f2f37586092df9f0e7e4b3d826d99f5d94812912fa92eee" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.172843 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.323781995 podStartE2EDuration="58.172827567s" podCreationTimestamp="2025-09-30 08:19:36 +0000 UTC" firstStartedPulling="2025-09-30 08:19:38.004717779 +0000 UTC m=+1001.456917046" lastFinishedPulling="2025-09-30 08:20:00.853763351 +0000 UTC m=+1024.305962618" observedRunningTime="2025-09-30 08:20:34.17052254 +0000 UTC m=+1057.622721807" watchObservedRunningTime="2025-09-30 08:20:34.172827567 +0000 UTC m=+1057.625026834" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.218341 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-notifications-server-0" podStartSLOduration=49.310430143 podStartE2EDuration="58.218317039s" podCreationTimestamp="2025-09-30 08:19:36 +0000 UTC" firstStartedPulling="2025-09-30 08:19:51.949980104 +0000 UTC m=+1015.402179371" lastFinishedPulling="2025-09-30 08:20:00.85786701 +0000 UTC m=+1024.310066267" observedRunningTime="2025-09-30 08:20:34.212038446 +0000 UTC m=+1057.664237703" watchObservedRunningTime="2025-09-30 08:20:34.218317039 +0000 UTC m=+1057.670516306" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.247117 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-777c66bf5c-8p229"] Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.259698 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-777c66bf5c-8p229"] Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.396963 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.499214 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvtwb\" (UniqueName: \"kubernetes.io/projected/8c9dead4-5ad1-40c7-8f76-2ebed2449991-kube-api-access-zvtwb\") pod \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.499260 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-dns-svc\") pod \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.499879 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-config\") pod \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.499902 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-ovsdbserver-nb\") pod \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\" (UID: \"8c9dead4-5ad1-40c7-8f76-2ebed2449991\") " Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.510285 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c9dead4-5ad1-40c7-8f76-2ebed2449991-kube-api-access-zvtwb" (OuterVolumeSpecName: "kube-api-access-zvtwb") pod "8c9dead4-5ad1-40c7-8f76-2ebed2449991" (UID: "8c9dead4-5ad1-40c7-8f76-2ebed2449991"). InnerVolumeSpecName "kube-api-access-zvtwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.529773 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8c9dead4-5ad1-40c7-8f76-2ebed2449991" (UID: "8c9dead4-5ad1-40c7-8f76-2ebed2449991"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.532189 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8c9dead4-5ad1-40c7-8f76-2ebed2449991" (UID: "8c9dead4-5ad1-40c7-8f76-2ebed2449991"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.539993 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-config" (OuterVolumeSpecName: "config") pod "8c9dead4-5ad1-40c7-8f76-2ebed2449991" (UID: "8c9dead4-5ad1-40c7-8f76-2ebed2449991"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.602505 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvtwb\" (UniqueName: \"kubernetes.io/projected/8c9dead4-5ad1-40c7-8f76-2ebed2449991-kube-api-access-zvtwb\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.602531 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.602539 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:34 crc kubenswrapper[4810]: I0930 08:20:34.602547 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c9dead4-5ad1-40c7-8f76-2ebed2449991-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:35 crc kubenswrapper[4810]: I0930 08:20:35.084909 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" event={"ID":"8c9dead4-5ad1-40c7-8f76-2ebed2449991","Type":"ContainerDied","Data":"469bacefc378a02c0911eeee1e28332be374cce8ffedfb816b15f091fb0954ec"} Sep 30 08:20:35 crc kubenswrapper[4810]: I0930 08:20:35.084974 4810 scope.go:117] "RemoveContainer" containerID="8b9d2848bf9b9b3622109d67971a004666e07a2906702ee98e963fc3c31bc76f" Sep 30 08:20:35 crc kubenswrapper[4810]: I0930 08:20:35.085080 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dbfbd97-tsh5g" Sep 30 08:20:35 crc kubenswrapper[4810]: I0930 08:20:35.090043 4810 generic.go:334] "Generic (PLEG): container finished" podID="73deb073-9ff6-4ef6-8e08-98ecc936dca1" containerID="fb2fc73b40d7f34790ae4b74b80e5d66fd77a31ddde5e3c943915c0bd78f339e" exitCode=0 Sep 30 08:20:35 crc kubenswrapper[4810]: I0930 08:20:35.090129 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" event={"ID":"73deb073-9ff6-4ef6-8e08-98ecc936dca1","Type":"ContainerDied","Data":"fb2fc73b40d7f34790ae4b74b80e5d66fd77a31ddde5e3c943915c0bd78f339e"} Sep 30 08:20:35 crc kubenswrapper[4810]: I0930 08:20:35.096003 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6wrgl" event={"ID":"4b9a3337-b2c2-49c8-9755-81660bb506af","Type":"ContainerStarted","Data":"920da9a81ae087730ac3871f47d47819203eb38889ccd8f81faa5f13e62db43b"} Sep 30 08:20:35 crc kubenswrapper[4810]: I0930 08:20:35.338997 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08ffde2b-6c3c-499f-b976-5aec9110596a" path="/var/lib/kubelet/pods/08ffde2b-6c3c-499f-b976-5aec9110596a/volumes" Sep 30 08:20:35 crc kubenswrapper[4810]: I0930 08:20:35.390894 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-6wrgl" podStartSLOduration=5.390834548 podStartE2EDuration="5.390834548s" podCreationTimestamp="2025-09-30 08:20:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:20:35.37713348 +0000 UTC m=+1058.829332747" watchObservedRunningTime="2025-09-30 08:20:35.390834548 +0000 UTC m=+1058.843033845" Sep 30 08:20:35 crc kubenswrapper[4810]: I0930 08:20:35.431288 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79dbfbd97-tsh5g"] Sep 30 08:20:35 crc kubenswrapper[4810]: I0930 08:20:35.437753 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79dbfbd97-tsh5g"] Sep 30 08:20:37 crc kubenswrapper[4810]: I0930 08:20:37.127682 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" event={"ID":"73deb073-9ff6-4ef6-8e08-98ecc936dca1","Type":"ContainerStarted","Data":"eba10c72387b210fe793bfada9c89f53eecbcaeb43a88bbfd9ccd9eaa76376d3"} Sep 30 08:20:37 crc kubenswrapper[4810]: I0930 08:20:37.128214 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:37 crc kubenswrapper[4810]: I0930 08:20:37.153652 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" podStartSLOduration=7.15363488 podStartE2EDuration="7.15363488s" podCreationTimestamp="2025-09-30 08:20:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:20:37.153450954 +0000 UTC m=+1060.605650221" watchObservedRunningTime="2025-09-30 08:20:37.15363488 +0000 UTC m=+1060.605834147" Sep 30 08:20:37 crc kubenswrapper[4810]: I0930 08:20:37.325861 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c9dead4-5ad1-40c7-8f76-2ebed2449991" path="/var/lib/kubelet/pods/8c9dead4-5ad1-40c7-8f76-2ebed2449991/volumes" Sep 30 08:20:38 crc kubenswrapper[4810]: I0930 08:20:38.597408 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-777c66bf5c-8p229" podUID="08ffde2b-6c3c-499f-b976-5aec9110596a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.120:5353: i/o timeout" Sep 30 08:20:40 crc kubenswrapper[4810]: I0930 08:20:40.156087 4810 generic.go:334] "Generic (PLEG): container finished" podID="7cb4aa97-f850-4e7f-a1e1-046bd5235109" containerID="faf03dfdbd92a5ec947277ac47ffcc4ccdbec7f33da4744535bbf55ed77a1bfd" exitCode=0 Sep 30 08:20:40 crc kubenswrapper[4810]: I0930 08:20:40.156221 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cb4aa97-f850-4e7f-a1e1-046bd5235109","Type":"ContainerDied","Data":"faf03dfdbd92a5ec947277ac47ffcc4ccdbec7f33da4744535bbf55ed77a1bfd"} Sep 30 08:20:40 crc kubenswrapper[4810]: I0930 08:20:40.221059 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:40 crc kubenswrapper[4810]: E0930 08:20:40.221554 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 08:20:40 crc kubenswrapper[4810]: E0930 08:20:40.221599 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 08:20:40 crc kubenswrapper[4810]: E0930 08:20:40.221686 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift podName:edd67220-e97a-465e-9671-2cfcf760475a nodeName:}" failed. No retries permitted until 2025-09-30 08:20:56.221661647 +0000 UTC m=+1079.673860984 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift") pod "swift-storage-0" (UID: "edd67220-e97a-465e-9671-2cfcf760475a") : configmap "swift-ring-files" not found Sep 30 08:20:41 crc kubenswrapper[4810]: I0930 08:20:41.061470 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:20:41 crc kubenswrapper[4810]: I0930 08:20:41.160696 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74ccbc479c-c75jf"] Sep 30 08:20:41 crc kubenswrapper[4810]: I0930 08:20:41.161244 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" podUID="7601717f-f941-49aa-bb26-309cfeb9401d" containerName="dnsmasq-dns" containerID="cri-o://586940b180c05e469c9e8b328fd16fe6ac2373eb3683b3bbe75d47cf9024eacc" gracePeriod=10 Sep 30 08:20:41 crc kubenswrapper[4810]: I0930 08:20:41.169381 4810 generic.go:334] "Generic (PLEG): container finished" podID="3ca490c3-ff2f-4553-9e66-4fb456d3a1f2" containerID="700445b48f9752d8bf4affa8a82fbcd516f7c05c7e3292183d4d078dfd37c443" exitCode=0 Sep 30 08:20:41 crc kubenswrapper[4810]: I0930 08:20:41.169435 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2","Type":"ContainerDied","Data":"700445b48f9752d8bf4affa8a82fbcd516f7c05c7e3292183d4d078dfd37c443"} Sep 30 08:20:41 crc kubenswrapper[4810]: I0930 08:20:41.753524 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" podUID="7601717f-f941-49aa-bb26-309cfeb9401d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.107:5353: connect: connection refused" Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.183487 4810 generic.go:334] "Generic (PLEG): container finished" podID="7601717f-f941-49aa-bb26-309cfeb9401d" containerID="586940b180c05e469c9e8b328fd16fe6ac2373eb3683b3bbe75d47cf9024eacc" exitCode=0 Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.183544 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" event={"ID":"7601717f-f941-49aa-bb26-309cfeb9401d","Type":"ContainerDied","Data":"586940b180c05e469c9e8b328fd16fe6ac2373eb3683b3bbe75d47cf9024eacc"} Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.499464 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-9q2d8" podUID="af34271b-1a44-4240-88ee-b86490744353" containerName="ovn-controller" probeResult="failure" output=< Sep 30 08:20:42 crc kubenswrapper[4810]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 30 08:20:42 crc kubenswrapper[4810]: > Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.749734 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.798349 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-dns-svc\") pod \"7601717f-f941-49aa-bb26-309cfeb9401d\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.798398 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-config\") pod \"7601717f-f941-49aa-bb26-309cfeb9401d\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.798463 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xkjn\" (UniqueName: \"kubernetes.io/projected/7601717f-f941-49aa-bb26-309cfeb9401d-kube-api-access-4xkjn\") pod \"7601717f-f941-49aa-bb26-309cfeb9401d\" (UID: \"7601717f-f941-49aa-bb26-309cfeb9401d\") " Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.815739 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7601717f-f941-49aa-bb26-309cfeb9401d-kube-api-access-4xkjn" (OuterVolumeSpecName: "kube-api-access-4xkjn") pod "7601717f-f941-49aa-bb26-309cfeb9401d" (UID: "7601717f-f941-49aa-bb26-309cfeb9401d"). InnerVolumeSpecName "kube-api-access-4xkjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.887913 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-config" (OuterVolumeSpecName: "config") pod "7601717f-f941-49aa-bb26-309cfeb9401d" (UID: "7601717f-f941-49aa-bb26-309cfeb9401d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.889184 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7601717f-f941-49aa-bb26-309cfeb9401d" (UID: "7601717f-f941-49aa-bb26-309cfeb9401d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.900691 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.900725 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7601717f-f941-49aa-bb26-309cfeb9401d-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:42 crc kubenswrapper[4810]: I0930 08:20:42.900740 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xkjn\" (UniqueName: \"kubernetes.io/projected/7601717f-f941-49aa-bb26-309cfeb9401d-kube-api-access-4xkjn\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.194508 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"306e4bf0-015f-4c7a-ba49-efda86e72442","Type":"ContainerStarted","Data":"c54010f7ebadc5c641d3920b807f1ea2e10fcfb60d29fd695e7dab5810e785a6"} Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.194565 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"306e4bf0-015f-4c7a-ba49-efda86e72442","Type":"ContainerStarted","Data":"edfe239f0f048a209862e1cfa42e481806ed788ff69781ca9aa72af9149b542c"} Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.194656 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.196389 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3ca490c3-ff2f-4553-9e66-4fb456d3a1f2","Type":"ContainerStarted","Data":"39682d3e7a762c52483dc422356e95ecf86367dad5939f3a80338d99a2b0c9e7"} Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.198098 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cb4aa97-f850-4e7f-a1e1-046bd5235109","Type":"ContainerStarted","Data":"132c6e91febb6fdc15c87bdc7e7cea6e82dbbbd4dd12ec86e93d4d768631745a"} Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.200144 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e9cd2658-4276-4957-9941-ab397d3e0d5f","Type":"ContainerStarted","Data":"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e"} Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.203870 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" event={"ID":"7601717f-f941-49aa-bb26-309cfeb9401d","Type":"ContainerDied","Data":"8d8758a484d8872511786184de235b5de21f1b47ca1977cf385e3e7367cdc501"} Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.203921 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74ccbc479c-c75jf" Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.203928 4810 scope.go:117] "RemoveContainer" containerID="586940b180c05e469c9e8b328fd16fe6ac2373eb3683b3bbe75d47cf9024eacc" Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.219872 4810 scope.go:117] "RemoveContainer" containerID="bc0f55a44eb1d2a6c99a1c6ae2c568445b637f08d7adbb0db8fee6ff072a319a" Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.263096 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.428458206 podStartE2EDuration="12.263064661s" podCreationTimestamp="2025-09-30 08:20:31 +0000 UTC" firstStartedPulling="2025-09-30 08:20:33.591856686 +0000 UTC m=+1057.044055953" lastFinishedPulling="2025-09-30 08:20:42.426463141 +0000 UTC m=+1065.878662408" observedRunningTime="2025-09-30 08:20:43.236179049 +0000 UTC m=+1066.688378306" watchObservedRunningTime="2025-09-30 08:20:43.263064661 +0000 UTC m=+1066.715263938" Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.265813 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=57.268938582 podStartE2EDuration="1m4.26580685s" podCreationTimestamp="2025-09-30 08:19:39 +0000 UTC" firstStartedPulling="2025-09-30 08:20:01.724304516 +0000 UTC m=+1025.176503793" lastFinishedPulling="2025-09-30 08:20:08.721172784 +0000 UTC m=+1032.173372061" observedRunningTime="2025-09-30 08:20:43.259922509 +0000 UTC m=+1066.712121786" watchObservedRunningTime="2025-09-30 08:20:43.26580685 +0000 UTC m=+1066.718006127" Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.300357 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=57.263849675 podStartE2EDuration="1m4.300338584s" podCreationTimestamp="2025-09-30 08:19:39 +0000 UTC" firstStartedPulling="2025-09-30 08:20:01.684688325 +0000 UTC m=+1025.136887592" lastFinishedPulling="2025-09-30 08:20:08.721177234 +0000 UTC m=+1032.173376501" observedRunningTime="2025-09-30 08:20:43.292105844 +0000 UTC m=+1066.744305131" watchObservedRunningTime="2025-09-30 08:20:43.300338584 +0000 UTC m=+1066.752537841" Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.322558 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74ccbc479c-c75jf"] Sep 30 08:20:43 crc kubenswrapper[4810]: I0930 08:20:43.322601 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74ccbc479c-c75jf"] Sep 30 08:20:44 crc kubenswrapper[4810]: I0930 08:20:44.228500 4810 generic.go:334] "Generic (PLEG): container finished" podID="cf76de28-c7b8-474f-93a7-5af536e817e8" containerID="98977281bf04b6773e81a3868aa50e6c5604dcb10433b7c39cf6f1c9906e4585" exitCode=0 Sep 30 08:20:44 crc kubenswrapper[4810]: I0930 08:20:44.228598 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zr8dp" event={"ID":"cf76de28-c7b8-474f-93a7-5af536e817e8","Type":"ContainerDied","Data":"98977281bf04b6773e81a3868aa50e6c5604dcb10433b7c39cf6f1c9906e4585"} Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.316241 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7601717f-f941-49aa-bb26-309cfeb9401d" path="/var/lib/kubelet/pods/7601717f-f941-49aa-bb26-309cfeb9401d/volumes" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.668504 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.753514 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-dispersionconf\") pod \"cf76de28-c7b8-474f-93a7-5af536e817e8\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.753638 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf76de28-c7b8-474f-93a7-5af536e817e8-etc-swift\") pod \"cf76de28-c7b8-474f-93a7-5af536e817e8\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.753809 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-swiftconf\") pod \"cf76de28-c7b8-474f-93a7-5af536e817e8\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.753857 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-scripts\") pod \"cf76de28-c7b8-474f-93a7-5af536e817e8\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.753924 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99mc9\" (UniqueName: \"kubernetes.io/projected/cf76de28-c7b8-474f-93a7-5af536e817e8-kube-api-access-99mc9\") pod \"cf76de28-c7b8-474f-93a7-5af536e817e8\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.753981 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-ring-data-devices\") pod \"cf76de28-c7b8-474f-93a7-5af536e817e8\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.754048 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-combined-ca-bundle\") pod \"cf76de28-c7b8-474f-93a7-5af536e817e8\" (UID: \"cf76de28-c7b8-474f-93a7-5af536e817e8\") " Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.754992 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf76de28-c7b8-474f-93a7-5af536e817e8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "cf76de28-c7b8-474f-93a7-5af536e817e8" (UID: "cf76de28-c7b8-474f-93a7-5af536e817e8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.755105 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "cf76de28-c7b8-474f-93a7-5af536e817e8" (UID: "cf76de28-c7b8-474f-93a7-5af536e817e8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.760294 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf76de28-c7b8-474f-93a7-5af536e817e8-kube-api-access-99mc9" (OuterVolumeSpecName: "kube-api-access-99mc9") pod "cf76de28-c7b8-474f-93a7-5af536e817e8" (UID: "cf76de28-c7b8-474f-93a7-5af536e817e8"). InnerVolumeSpecName "kube-api-access-99mc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.771169 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "cf76de28-c7b8-474f-93a7-5af536e817e8" (UID: "cf76de28-c7b8-474f-93a7-5af536e817e8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.795289 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "cf76de28-c7b8-474f-93a7-5af536e817e8" (UID: "cf76de28-c7b8-474f-93a7-5af536e817e8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.797757 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-scripts" (OuterVolumeSpecName: "scripts") pod "cf76de28-c7b8-474f-93a7-5af536e817e8" (UID: "cf76de28-c7b8-474f-93a7-5af536e817e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.819346 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf76de28-c7b8-474f-93a7-5af536e817e8" (UID: "cf76de28-c7b8-474f-93a7-5af536e817e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.857072 4810 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-swiftconf\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.857128 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.857154 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99mc9\" (UniqueName: \"kubernetes.io/projected/cf76de28-c7b8-474f-93a7-5af536e817e8-kube-api-access-99mc9\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.857179 4810 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf76de28-c7b8-474f-93a7-5af536e817e8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.857198 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.857216 4810 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf76de28-c7b8-474f-93a7-5af536e817e8-dispersionconf\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.857233 4810 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf76de28-c7b8-474f-93a7-5af536e817e8-etc-swift\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.911452 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.912326 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.912525 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.913658 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d1ca8f9a2402180e7c60714ae09e65a733987d89bb171e82c451bfe63cec7eed"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:20:45 crc kubenswrapper[4810]: I0930 08:20:45.913927 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://d1ca8f9a2402180e7c60714ae09e65a733987d89bb171e82c451bfe63cec7eed" gracePeriod=600 Sep 30 08:20:46 crc kubenswrapper[4810]: I0930 08:20:46.251708 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e9cd2658-4276-4957-9941-ab397d3e0d5f","Type":"ContainerStarted","Data":"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703"} Sep 30 08:20:46 crc kubenswrapper[4810]: I0930 08:20:46.253497 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zr8dp" event={"ID":"cf76de28-c7b8-474f-93a7-5af536e817e8","Type":"ContainerDied","Data":"03ee4defcb039cff19db4cf82a2c3165daa68373bd332d0eece308e9cb9fedb3"} Sep 30 08:20:46 crc kubenswrapper[4810]: I0930 08:20:46.253520 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03ee4defcb039cff19db4cf82a2c3165daa68373bd332d0eece308e9cb9fedb3" Sep 30 08:20:46 crc kubenswrapper[4810]: I0930 08:20:46.253702 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zr8dp" Sep 30 08:20:46 crc kubenswrapper[4810]: I0930 08:20:46.257047 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="d1ca8f9a2402180e7c60714ae09e65a733987d89bb171e82c451bfe63cec7eed" exitCode=0 Sep 30 08:20:46 crc kubenswrapper[4810]: I0930 08:20:46.257081 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"d1ca8f9a2402180e7c60714ae09e65a733987d89bb171e82c451bfe63cec7eed"} Sep 30 08:20:46 crc kubenswrapper[4810]: I0930 08:20:46.257176 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"d77e6c1674e48be39d628cea92dce547143c707ab6bb630e41eb5a3542975559"} Sep 30 08:20:46 crc kubenswrapper[4810]: I0930 08:20:46.257208 4810 scope.go:117] "RemoveContainer" containerID="b2b0b0f4ca79c884d37aac25744a2d16466a6c0fbe29d05bbf5e9792b146f420" Sep 30 08:20:47 crc kubenswrapper[4810]: I0930 08:20:47.139833 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="5bb3b225-5221-417f-9914-737723ac52aa" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Sep 30 08:20:47 crc kubenswrapper[4810]: I0930 08:20:47.466570 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="cf638129-9405-48c6-8458-143b297363fd" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Sep 30 08:20:47 crc kubenswrapper[4810]: I0930 08:20:47.480023 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-9q2d8" podUID="af34271b-1a44-4240-88ee-b86490744353" containerName="ovn-controller" probeResult="failure" output=< Sep 30 08:20:47 crc kubenswrapper[4810]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 30 08:20:47 crc kubenswrapper[4810]: > Sep 30 08:20:47 crc kubenswrapper[4810]: I0930 08:20:47.835889 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-notifications-server-0" podUID="fcf1eb6a-de6d-48ac-bbf5-4e015c452e18" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.110:5671: connect: connection refused" Sep 30 08:20:50 crc kubenswrapper[4810]: I0930 08:20:50.309916 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e9cd2658-4276-4957-9941-ab397d3e0d5f","Type":"ContainerStarted","Data":"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a"} Sep 30 08:20:50 crc kubenswrapper[4810]: I0930 08:20:50.354872 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=19.630925789 podStartE2EDuration="1m7.354842405s" podCreationTimestamp="2025-09-30 08:19:43 +0000 UTC" firstStartedPulling="2025-09-30 08:20:01.442074095 +0000 UTC m=+1024.894273362" lastFinishedPulling="2025-09-30 08:20:49.165990691 +0000 UTC m=+1072.618189978" observedRunningTime="2025-09-30 08:20:50.343962869 +0000 UTC m=+1073.796162176" watchObservedRunningTime="2025-09-30 08:20:50.354842405 +0000 UTC m=+1073.807041682" Sep 30 08:20:50 crc kubenswrapper[4810]: I0930 08:20:50.762048 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Sep 30 08:20:50 crc kubenswrapper[4810]: I0930 08:20:50.762497 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Sep 30 08:20:50 crc kubenswrapper[4810]: I0930 08:20:50.901203 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Sep 30 08:20:51 crc kubenswrapper[4810]: I0930 08:20:51.144755 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Sep 30 08:20:51 crc kubenswrapper[4810]: I0930 08:20:51.145142 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Sep 30 08:20:51 crc kubenswrapper[4810]: I0930 08:20:51.238002 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Sep 30 08:20:51 crc kubenswrapper[4810]: I0930 08:20:51.404010 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Sep 30 08:20:51 crc kubenswrapper[4810]: I0930 08:20:51.411348 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Sep 30 08:20:52 crc kubenswrapper[4810]: I0930 08:20:52.478476 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-9q2d8" podUID="af34271b-1a44-4240-88ee-b86490744353" containerName="ovn-controller" probeResult="failure" output=< Sep 30 08:20:52 crc kubenswrapper[4810]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 30 08:20:52 crc kubenswrapper[4810]: > Sep 30 08:20:52 crc kubenswrapper[4810]: I0930 08:20:52.510771 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.241496 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-h844r"] Sep 30 08:20:53 crc kubenswrapper[4810]: E0930 08:20:53.241893 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf76de28-c7b8-474f-93a7-5af536e817e8" containerName="swift-ring-rebalance" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.241954 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf76de28-c7b8-474f-93a7-5af536e817e8" containerName="swift-ring-rebalance" Sep 30 08:20:53 crc kubenswrapper[4810]: E0930 08:20:53.241969 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ffde2b-6c3c-499f-b976-5aec9110596a" containerName="init" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.241977 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ffde2b-6c3c-499f-b976-5aec9110596a" containerName="init" Sep 30 08:20:53 crc kubenswrapper[4810]: E0930 08:20:53.241997 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ffde2b-6c3c-499f-b976-5aec9110596a" containerName="dnsmasq-dns" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.242006 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ffde2b-6c3c-499f-b976-5aec9110596a" containerName="dnsmasq-dns" Sep 30 08:20:53 crc kubenswrapper[4810]: E0930 08:20:53.242029 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7601717f-f941-49aa-bb26-309cfeb9401d" containerName="init" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.242037 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7601717f-f941-49aa-bb26-309cfeb9401d" containerName="init" Sep 30 08:20:53 crc kubenswrapper[4810]: E0930 08:20:53.242050 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9dead4-5ad1-40c7-8f76-2ebed2449991" containerName="init" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.242061 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9dead4-5ad1-40c7-8f76-2ebed2449991" containerName="init" Sep 30 08:20:53 crc kubenswrapper[4810]: E0930 08:20:53.242077 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7601717f-f941-49aa-bb26-309cfeb9401d" containerName="dnsmasq-dns" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.242095 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7601717f-f941-49aa-bb26-309cfeb9401d" containerName="dnsmasq-dns" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.242338 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf76de28-c7b8-474f-93a7-5af536e817e8" containerName="swift-ring-rebalance" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.242363 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7601717f-f941-49aa-bb26-309cfeb9401d" containerName="dnsmasq-dns" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.242382 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="08ffde2b-6c3c-499f-b976-5aec9110596a" containerName="dnsmasq-dns" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.242396 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9dead4-5ad1-40c7-8f76-2ebed2449991" containerName="init" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.243047 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-h844r" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.251812 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-h844r"] Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.298825 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qht9b\" (UniqueName: \"kubernetes.io/projected/7dacf67c-34b5-4c58-9f8d-00adfdcf74b7-kube-api-access-qht9b\") pod \"watcher-db-create-h844r\" (UID: \"7dacf67c-34b5-4c58-9f8d-00adfdcf74b7\") " pod="openstack/watcher-db-create-h844r" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.401074 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qht9b\" (UniqueName: \"kubernetes.io/projected/7dacf67c-34b5-4c58-9f8d-00adfdcf74b7-kube-api-access-qht9b\") pod \"watcher-db-create-h844r\" (UID: \"7dacf67c-34b5-4c58-9f8d-00adfdcf74b7\") " pod="openstack/watcher-db-create-h844r" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.426951 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qht9b\" (UniqueName: \"kubernetes.io/projected/7dacf67c-34b5-4c58-9f8d-00adfdcf74b7-kube-api-access-qht9b\") pod \"watcher-db-create-h844r\" (UID: \"7dacf67c-34b5-4c58-9f8d-00adfdcf74b7\") " pod="openstack/watcher-db-create-h844r" Sep 30 08:20:53 crc kubenswrapper[4810]: I0930 08:20:53.574295 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-h844r" Sep 30 08:20:54 crc kubenswrapper[4810]: I0930 08:20:54.080205 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-h844r"] Sep 30 08:20:54 crc kubenswrapper[4810]: W0930 08:20:54.088043 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dacf67c_34b5_4c58_9f8d_00adfdcf74b7.slice/crio-b51099892c3497b33dce1e8b3792cd4ae759cb9e7b40a3543c144dbe6b5a3499 WatchSource:0}: Error finding container b51099892c3497b33dce1e8b3792cd4ae759cb9e7b40a3543c144dbe6b5a3499: Status 404 returned error can't find the container with id b51099892c3497b33dce1e8b3792cd4ae759cb9e7b40a3543c144dbe6b5a3499 Sep 30 08:20:54 crc kubenswrapper[4810]: I0930 08:20:54.363433 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-h844r" event={"ID":"7dacf67c-34b5-4c58-9f8d-00adfdcf74b7","Type":"ContainerStarted","Data":"3d34db2a649ce680a9f15f98539d7a3bf4ba3020b2242e6b1eb8a4bec6e3bf54"} Sep 30 08:20:54 crc kubenswrapper[4810]: I0930 08:20:54.363838 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-h844r" event={"ID":"7dacf67c-34b5-4c58-9f8d-00adfdcf74b7","Type":"ContainerStarted","Data":"b51099892c3497b33dce1e8b3792cd4ae759cb9e7b40a3543c144dbe6b5a3499"} Sep 30 08:20:54 crc kubenswrapper[4810]: I0930 08:20:54.385556 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-create-h844r" podStartSLOduration=1.385536925 podStartE2EDuration="1.385536925s" podCreationTimestamp="2025-09-30 08:20:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:20:54.382742353 +0000 UTC m=+1077.834941700" watchObservedRunningTime="2025-09-30 08:20:54.385536925 +0000 UTC m=+1077.837736192" Sep 30 08:20:54 crc kubenswrapper[4810]: I0930 08:20:54.475873 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Sep 30 08:20:54 crc kubenswrapper[4810]: E0930 08:20:54.500531 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dacf67c_34b5_4c58_9f8d_00adfdcf74b7.slice/crio-conmon-3d34db2a649ce680a9f15f98539d7a3bf4ba3020b2242e6b1eb8a4bec6e3bf54.scope\": RecentStats: unable to find data in memory cache]" Sep 30 08:20:55 crc kubenswrapper[4810]: I0930 08:20:55.377011 4810 generic.go:334] "Generic (PLEG): container finished" podID="7dacf67c-34b5-4c58-9f8d-00adfdcf74b7" containerID="3d34db2a649ce680a9f15f98539d7a3bf4ba3020b2242e6b1eb8a4bec6e3bf54" exitCode=0 Sep 30 08:20:55 crc kubenswrapper[4810]: I0930 08:20:55.377053 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-h844r" event={"ID":"7dacf67c-34b5-4c58-9f8d-00adfdcf74b7","Type":"ContainerDied","Data":"3d34db2a649ce680a9f15f98539d7a3bf4ba3020b2242e6b1eb8a4bec6e3bf54"} Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.261000 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.273323 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd67220-e97a-465e-9671-2cfcf760475a-etc-swift\") pod \"swift-storage-0\" (UID: \"edd67220-e97a-465e-9671-2cfcf760475a\") " pod="openstack/swift-storage-0" Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.492963 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.556826 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Sep 30 08:20:56 crc kubenswrapper[4810]: E0930 08:20:56.616808 4810 kubelet_node_status.go:756] "Failed to set some node status fields" err="failed to validate nodeIP: route ip+net: no such network interface" node="crc" Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.761644 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-xqn7n"] Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.763082 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xqn7n" Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.767384 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xqn7n"] Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.773464 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz7bq\" (UniqueName: \"kubernetes.io/projected/531ec0c6-aa15-4799-b0a7-594fa996f069-kube-api-access-nz7bq\") pod \"glance-db-create-xqn7n\" (UID: \"531ec0c6-aa15-4799-b0a7-594fa996f069\") " pod="openstack/glance-db-create-xqn7n" Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.792399 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-h844r" Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.875989 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz7bq\" (UniqueName: \"kubernetes.io/projected/531ec0c6-aa15-4799-b0a7-594fa996f069-kube-api-access-nz7bq\") pod \"glance-db-create-xqn7n\" (UID: \"531ec0c6-aa15-4799-b0a7-594fa996f069\") " pod="openstack/glance-db-create-xqn7n" Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.896810 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz7bq\" (UniqueName: \"kubernetes.io/projected/531ec0c6-aa15-4799-b0a7-594fa996f069-kube-api-access-nz7bq\") pod \"glance-db-create-xqn7n\" (UID: \"531ec0c6-aa15-4799-b0a7-594fa996f069\") " pod="openstack/glance-db-create-xqn7n" Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.977597 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qht9b\" (UniqueName: \"kubernetes.io/projected/7dacf67c-34b5-4c58-9f8d-00adfdcf74b7-kube-api-access-qht9b\") pod \"7dacf67c-34b5-4c58-9f8d-00adfdcf74b7\" (UID: \"7dacf67c-34b5-4c58-9f8d-00adfdcf74b7\") " Sep 30 08:20:56 crc kubenswrapper[4810]: I0930 08:20:56.981313 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dacf67c-34b5-4c58-9f8d-00adfdcf74b7-kube-api-access-qht9b" (OuterVolumeSpecName: "kube-api-access-qht9b") pod "7dacf67c-34b5-4c58-9f8d-00adfdcf74b7" (UID: "7dacf67c-34b5-4c58-9f8d-00adfdcf74b7"). InnerVolumeSpecName "kube-api-access-qht9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.079670 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qht9b\" (UniqueName: \"kubernetes.io/projected/7dacf67c-34b5-4c58-9f8d-00adfdcf74b7-kube-api-access-qht9b\") on node \"crc\" DevicePath \"\"" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.108662 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xqn7n" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.138472 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.151456 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 30 08:20:57 crc kubenswrapper[4810]: W0930 08:20:57.155715 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedd67220_e97a_465e_9671_2cfcf760475a.slice/crio-f6ffcdf22f46f0d1104a0d89c8f57d810dc820980335b7e7e35d4a771db375df WatchSource:0}: Error finding container f6ffcdf22f46f0d1104a0d89c8f57d810dc820980335b7e7e35d4a771db375df: Status 404 returned error can't find the container with id f6ffcdf22f46f0d1104a0d89c8f57d810dc820980335b7e7e35d4a771db375df Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.576223 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.601544 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-jgr2f"] Sep 30 08:20:57 crc kubenswrapper[4810]: E0930 08:20:57.602313 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dacf67c-34b5-4c58-9f8d-00adfdcf74b7" containerName="mariadb-database-create" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.602380 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dacf67c-34b5-4c58-9f8d-00adfdcf74b7" containerName="mariadb-database-create" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.602639 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dacf67c-34b5-4c58-9f8d-00adfdcf74b7" containerName="mariadb-database-create" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.603352 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jgr2f" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.612543 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jgr2f"] Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.616487 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-h844r" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.616910 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-h844r" event={"ID":"7dacf67c-34b5-4c58-9f8d-00adfdcf74b7","Type":"ContainerDied","Data":"b51099892c3497b33dce1e8b3792cd4ae759cb9e7b40a3543c144dbe6b5a3499"} Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.617030 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b51099892c3497b33dce1e8b3792cd4ae759cb9e7b40a3543c144dbe6b5a3499" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.632714 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"f6ffcdf22f46f0d1104a0d89c8f57d810dc820980335b7e7e35d4a771db375df"} Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.646077 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xqn7n"] Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.651512 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-9q2d8" podUID="af34271b-1a44-4240-88ee-b86490744353" containerName="ovn-controller" probeResult="failure" output=< Sep 30 08:20:57 crc kubenswrapper[4810]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 30 08:20:57 crc kubenswrapper[4810]: > Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.677628 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mt698" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.681201 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-tdcfg"] Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.684603 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tdcfg" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.690708 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-tdcfg"] Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.783424 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpb7p\" (UniqueName: \"kubernetes.io/projected/fef53718-de39-49b8-a9af-ef096a5e88ed-kube-api-access-tpb7p\") pod \"cinder-db-create-jgr2f\" (UID: \"fef53718-de39-49b8-a9af-ef096a5e88ed\") " pod="openstack/cinder-db-create-jgr2f" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.783501 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpcsp\" (UniqueName: \"kubernetes.io/projected/295208bb-eebe-474d-9210-be3a67db28b0-kube-api-access-qpcsp\") pod \"barbican-db-create-tdcfg\" (UID: \"295208bb-eebe-474d-9210-be3a67db28b0\") " pod="openstack/barbican-db-create-tdcfg" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.836541 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-notifications-server-0" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.851867 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-95ds9"] Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.852916 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-95ds9" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.885718 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpb7p\" (UniqueName: \"kubernetes.io/projected/fef53718-de39-49b8-a9af-ef096a5e88ed-kube-api-access-tpb7p\") pod \"cinder-db-create-jgr2f\" (UID: \"fef53718-de39-49b8-a9af-ef096a5e88ed\") " pod="openstack/cinder-db-create-jgr2f" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.885803 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpcsp\" (UniqueName: \"kubernetes.io/projected/295208bb-eebe-474d-9210-be3a67db28b0-kube-api-access-qpcsp\") pod \"barbican-db-create-tdcfg\" (UID: \"295208bb-eebe-474d-9210-be3a67db28b0\") " pod="openstack/barbican-db-create-tdcfg" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.886859 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-95ds9"] Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.912236 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpb7p\" (UniqueName: \"kubernetes.io/projected/fef53718-de39-49b8-a9af-ef096a5e88ed-kube-api-access-tpb7p\") pod \"cinder-db-create-jgr2f\" (UID: \"fef53718-de39-49b8-a9af-ef096a5e88ed\") " pod="openstack/cinder-db-create-jgr2f" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.921993 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpcsp\" (UniqueName: \"kubernetes.io/projected/295208bb-eebe-474d-9210-be3a67db28b0-kube-api-access-qpcsp\") pod \"barbican-db-create-tdcfg\" (UID: \"295208bb-eebe-474d-9210-be3a67db28b0\") " pod="openstack/barbican-db-create-tdcfg" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.976515 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9q2d8-config-x5ckn"] Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.980034 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jgr2f" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.989459 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.990251 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fctpm\" (UniqueName: \"kubernetes.io/projected/af23fdb8-4a62-427f-8197-ca7a551b7b22-kube-api-access-fctpm\") pod \"neutron-db-create-95ds9\" (UID: \"af23fdb8-4a62-427f-8197-ca7a551b7b22\") " pod="openstack/neutron-db-create-95ds9" Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.992938 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9q2d8-config-x5ckn"] Sep 30 08:20:57 crc kubenswrapper[4810]: I0930 08:20:57.993897 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.086250 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tdcfg" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.091855 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fctpm\" (UniqueName: \"kubernetes.io/projected/af23fdb8-4a62-427f-8197-ca7a551b7b22-kube-api-access-fctpm\") pod \"neutron-db-create-95ds9\" (UID: \"af23fdb8-4a62-427f-8197-ca7a551b7b22\") " pod="openstack/neutron-db-create-95ds9" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.092098 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zblvq\" (UniqueName: \"kubernetes.io/projected/12c190a8-686e-4984-886e-ab7fb72f3d0e-kube-api-access-zblvq\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.092235 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-additional-scripts\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.092359 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.092433 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run-ovn\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.092513 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-log-ovn\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.092604 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-scripts\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.126634 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fctpm\" (UniqueName: \"kubernetes.io/projected/af23fdb8-4a62-427f-8197-ca7a551b7b22-kube-api-access-fctpm\") pod \"neutron-db-create-95ds9\" (UID: \"af23fdb8-4a62-427f-8197-ca7a551b7b22\") " pod="openstack/neutron-db-create-95ds9" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.184518 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-95ds9" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.197393 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-additional-scripts\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.197460 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.197489 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run-ovn\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.197523 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-log-ovn\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.197579 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-scripts\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.197685 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zblvq\" (UniqueName: \"kubernetes.io/projected/12c190a8-686e-4984-886e-ab7fb72f3d0e-kube-api-access-zblvq\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.197804 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.197874 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-log-ovn\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.197922 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run-ovn\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.198143 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-additional-scripts\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.202144 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-scripts\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.225985 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zblvq\" (UniqueName: \"kubernetes.io/projected/12c190a8-686e-4984-886e-ab7fb72f3d0e-kube-api-access-zblvq\") pod \"ovn-controller-9q2d8-config-x5ckn\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.320035 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.463044 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jgr2f"] Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.644010 4810 generic.go:334] "Generic (PLEG): container finished" podID="531ec0c6-aa15-4799-b0a7-594fa996f069" containerID="d5506f9d5f727624787519b93d0b7fd28688f63ac436751c0bb78b6ce632b1b1" exitCode=0 Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.644067 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xqn7n" event={"ID":"531ec0c6-aa15-4799-b0a7-594fa996f069","Type":"ContainerDied","Data":"d5506f9d5f727624787519b93d0b7fd28688f63ac436751c0bb78b6ce632b1b1"} Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.644104 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xqn7n" event={"ID":"531ec0c6-aa15-4799-b0a7-594fa996f069","Type":"ContainerStarted","Data":"b78cbb67e88445a6b043b8986e8907a81a531ddb868c2ff076fd7e2a41e2ce65"} Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.734041 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-95ds9"] Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.741656 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-tdcfg"] Sep 30 08:20:58 crc kubenswrapper[4810]: W0930 08:20:58.811184 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf23fdb8_4a62_427f_8197_ca7a551b7b22.slice/crio-c1c94ed971921a329482c622028d3dbee0f5a5e437cd37435737b834920f7303 WatchSource:0}: Error finding container c1c94ed971921a329482c622028d3dbee0f5a5e437cd37435737b834920f7303: Status 404 returned error can't find the container with id c1c94ed971921a329482c622028d3dbee0f5a5e437cd37435737b834920f7303 Sep 30 08:20:58 crc kubenswrapper[4810]: E0930 08:20:58.884759 4810 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.203:50444->38.102.83.203:42869: write tcp 38.102.83.203:50444->38.102.83.203:42869: write: broken pipe Sep 30 08:20:58 crc kubenswrapper[4810]: I0930 08:20:58.890402 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9q2d8-config-x5ckn"] Sep 30 08:20:58 crc kubenswrapper[4810]: W0930 08:20:58.916603 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12c190a8_686e_4984_886e_ab7fb72f3d0e.slice/crio-4eb6fd3fce833bbde787e0a56aa307dbc5c513848400c0d8467bdc1a92e7ee8d WatchSource:0}: Error finding container 4eb6fd3fce833bbde787e0a56aa307dbc5c513848400c0d8467bdc1a92e7ee8d: Status 404 returned error can't find the container with id 4eb6fd3fce833bbde787e0a56aa307dbc5c513848400c0d8467bdc1a92e7ee8d Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.476892 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.479662 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.653823 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8-config-x5ckn" event={"ID":"12c190a8-686e-4984-886e-ab7fb72f3d0e","Type":"ContainerStarted","Data":"ae64ed2fa0a9c068aaa9bf2860a2b815a9b72f9685c9c6814961bf0848668bf1"} Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.654058 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8-config-x5ckn" event={"ID":"12c190a8-686e-4984-886e-ab7fb72f3d0e","Type":"ContainerStarted","Data":"4eb6fd3fce833bbde787e0a56aa307dbc5c513848400c0d8467bdc1a92e7ee8d"} Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.657103 4810 generic.go:334] "Generic (PLEG): container finished" podID="295208bb-eebe-474d-9210-be3a67db28b0" containerID="c9cf8eb6b3fd5cc334ba71d12ecc662c9cc0c7d77bf3a629147481625c1bf1d2" exitCode=0 Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.657133 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tdcfg" event={"ID":"295208bb-eebe-474d-9210-be3a67db28b0","Type":"ContainerDied","Data":"c9cf8eb6b3fd5cc334ba71d12ecc662c9cc0c7d77bf3a629147481625c1bf1d2"} Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.657171 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tdcfg" event={"ID":"295208bb-eebe-474d-9210-be3a67db28b0","Type":"ContainerStarted","Data":"3450f901264838c3bfbba200f42c0cb0e8daa70d7c985642ac5e4a6bad564adf"} Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.659018 4810 generic.go:334] "Generic (PLEG): container finished" podID="af23fdb8-4a62-427f-8197-ca7a551b7b22" containerID="508caefa1a2537eeb65f5e2cbea3da41cd51ebc16c43bf2a167622e5f83972b0" exitCode=0 Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.659075 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-95ds9" event={"ID":"af23fdb8-4a62-427f-8197-ca7a551b7b22","Type":"ContainerDied","Data":"508caefa1a2537eeb65f5e2cbea3da41cd51ebc16c43bf2a167622e5f83972b0"} Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.659097 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-95ds9" event={"ID":"af23fdb8-4a62-427f-8197-ca7a551b7b22","Type":"ContainerStarted","Data":"c1c94ed971921a329482c622028d3dbee0f5a5e437cd37435737b834920f7303"} Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.661119 4810 generic.go:334] "Generic (PLEG): container finished" podID="fef53718-de39-49b8-a9af-ef096a5e88ed" containerID="64ac1a88b3cc20b7584a4718f6cea16810f7a2767bcba16b6844af74548d9def" exitCode=0 Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.661209 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jgr2f" event={"ID":"fef53718-de39-49b8-a9af-ef096a5e88ed","Type":"ContainerDied","Data":"64ac1a88b3cc20b7584a4718f6cea16810f7a2767bcba16b6844af74548d9def"} Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.661279 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jgr2f" event={"ID":"fef53718-de39-49b8-a9af-ef096a5e88ed","Type":"ContainerStarted","Data":"47438bbd3f5226677ff0404ee936923f60f768df8ae733480702da235b9cc857"} Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.663872 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"2bb94c919fa7f837a5416ae984467df8e69a8236dcf8a3da4f2a89c96e08a7f5"} Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.663911 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"672b247ca5625163956697f8f9a06c2e3aabaa32fefe380f0a95ce87d4d364c4"} Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.665762 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Sep 30 08:20:59 crc kubenswrapper[4810]: I0930 08:20:59.682817 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-9q2d8-config-x5ckn" podStartSLOduration=2.6828029559999997 podStartE2EDuration="2.682802956s" podCreationTimestamp="2025-09-30 08:20:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:20:59.682024843 +0000 UTC m=+1083.134224110" watchObservedRunningTime="2025-09-30 08:20:59.682802956 +0000 UTC m=+1083.135002223" Sep 30 08:21:00 crc kubenswrapper[4810]: I0930 08:21:00.110905 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xqn7n" Sep 30 08:21:00 crc kubenswrapper[4810]: I0930 08:21:00.235814 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz7bq\" (UniqueName: \"kubernetes.io/projected/531ec0c6-aa15-4799-b0a7-594fa996f069-kube-api-access-nz7bq\") pod \"531ec0c6-aa15-4799-b0a7-594fa996f069\" (UID: \"531ec0c6-aa15-4799-b0a7-594fa996f069\") " Sep 30 08:21:00 crc kubenswrapper[4810]: I0930 08:21:00.246450 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/531ec0c6-aa15-4799-b0a7-594fa996f069-kube-api-access-nz7bq" (OuterVolumeSpecName: "kube-api-access-nz7bq") pod "531ec0c6-aa15-4799-b0a7-594fa996f069" (UID: "531ec0c6-aa15-4799-b0a7-594fa996f069"). InnerVolumeSpecName "kube-api-access-nz7bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:00 crc kubenswrapper[4810]: I0930 08:21:00.340197 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz7bq\" (UniqueName: \"kubernetes.io/projected/531ec0c6-aa15-4799-b0a7-594fa996f069-kube-api-access-nz7bq\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:00 crc kubenswrapper[4810]: I0930 08:21:00.676346 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"9cec648f9bf5b4f5d707b8a9fa369792ab814094ed11566b50cd3455d2b0c1ab"} Sep 30 08:21:00 crc kubenswrapper[4810]: I0930 08:21:00.676383 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"3f368d02b5817a4bba941f5b1e553761d9ac281b9a087a4889268e611c14480d"} Sep 30 08:21:00 crc kubenswrapper[4810]: I0930 08:21:00.679058 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xqn7n" event={"ID":"531ec0c6-aa15-4799-b0a7-594fa996f069","Type":"ContainerDied","Data":"b78cbb67e88445a6b043b8986e8907a81a531ddb868c2ff076fd7e2a41e2ce65"} Sep 30 08:21:00 crc kubenswrapper[4810]: I0930 08:21:00.679098 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b78cbb67e88445a6b043b8986e8907a81a531ddb868c2ff076fd7e2a41e2ce65" Sep 30 08:21:00 crc kubenswrapper[4810]: I0930 08:21:00.679128 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xqn7n" Sep 30 08:21:00 crc kubenswrapper[4810]: I0930 08:21:00.680928 4810 generic.go:334] "Generic (PLEG): container finished" podID="12c190a8-686e-4984-886e-ab7fb72f3d0e" containerID="ae64ed2fa0a9c068aaa9bf2860a2b815a9b72f9685c9c6814961bf0848668bf1" exitCode=0 Sep 30 08:21:00 crc kubenswrapper[4810]: I0930 08:21:00.681005 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8-config-x5ckn" event={"ID":"12c190a8-686e-4984-886e-ab7fb72f3d0e","Type":"ContainerDied","Data":"ae64ed2fa0a9c068aaa9bf2860a2b815a9b72f9685c9c6814961bf0848668bf1"} Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.175653 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-c7s2d"] Sep 30 08:21:01 crc kubenswrapper[4810]: E0930 08:21:01.176592 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="531ec0c6-aa15-4799-b0a7-594fa996f069" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.176622 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="531ec0c6-aa15-4799-b0a7-594fa996f069" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.176850 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="531ec0c6-aa15-4799-b0a7-594fa996f069" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.177627 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c7s2d" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.182745 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-c7s2d"] Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.257420 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5gj2\" (UniqueName: \"kubernetes.io/projected/7af7ec1f-cde8-4a71-953d-ed7426ba7727-kube-api-access-h5gj2\") pod \"keystone-db-create-c7s2d\" (UID: \"7af7ec1f-cde8-4a71-953d-ed7426ba7727\") " pod="openstack/keystone-db-create-c7s2d" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.279001 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-95ds9" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.289067 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jgr2f" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.324573 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tdcfg" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.358724 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpcsp\" (UniqueName: \"kubernetes.io/projected/295208bb-eebe-474d-9210-be3a67db28b0-kube-api-access-qpcsp\") pod \"295208bb-eebe-474d-9210-be3a67db28b0\" (UID: \"295208bb-eebe-474d-9210-be3a67db28b0\") " Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.358819 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fctpm\" (UniqueName: \"kubernetes.io/projected/af23fdb8-4a62-427f-8197-ca7a551b7b22-kube-api-access-fctpm\") pod \"af23fdb8-4a62-427f-8197-ca7a551b7b22\" (UID: \"af23fdb8-4a62-427f-8197-ca7a551b7b22\") " Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.358954 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpb7p\" (UniqueName: \"kubernetes.io/projected/fef53718-de39-49b8-a9af-ef096a5e88ed-kube-api-access-tpb7p\") pod \"fef53718-de39-49b8-a9af-ef096a5e88ed\" (UID: \"fef53718-de39-49b8-a9af-ef096a5e88ed\") " Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.359206 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5gj2\" (UniqueName: \"kubernetes.io/projected/7af7ec1f-cde8-4a71-953d-ed7426ba7727-kube-api-access-h5gj2\") pod \"keystone-db-create-c7s2d\" (UID: \"7af7ec1f-cde8-4a71-953d-ed7426ba7727\") " pod="openstack/keystone-db-create-c7s2d" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.364785 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fef53718-de39-49b8-a9af-ef096a5e88ed-kube-api-access-tpb7p" (OuterVolumeSpecName: "kube-api-access-tpb7p") pod "fef53718-de39-49b8-a9af-ef096a5e88ed" (UID: "fef53718-de39-49b8-a9af-ef096a5e88ed"). InnerVolumeSpecName "kube-api-access-tpb7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.365349 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/295208bb-eebe-474d-9210-be3a67db28b0-kube-api-access-qpcsp" (OuterVolumeSpecName: "kube-api-access-qpcsp") pod "295208bb-eebe-474d-9210-be3a67db28b0" (UID: "295208bb-eebe-474d-9210-be3a67db28b0"). InnerVolumeSpecName "kube-api-access-qpcsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.367456 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af23fdb8-4a62-427f-8197-ca7a551b7b22-kube-api-access-fctpm" (OuterVolumeSpecName: "kube-api-access-fctpm") pod "af23fdb8-4a62-427f-8197-ca7a551b7b22" (UID: "af23fdb8-4a62-427f-8197-ca7a551b7b22"). InnerVolumeSpecName "kube-api-access-fctpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.376439 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5gj2\" (UniqueName: \"kubernetes.io/projected/7af7ec1f-cde8-4a71-953d-ed7426ba7727-kube-api-access-h5gj2\") pod \"keystone-db-create-c7s2d\" (UID: \"7af7ec1f-cde8-4a71-953d-ed7426ba7727\") " pod="openstack/keystone-db-create-c7s2d" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.460455 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fctpm\" (UniqueName: \"kubernetes.io/projected/af23fdb8-4a62-427f-8197-ca7a551b7b22-kube-api-access-fctpm\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.460489 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpb7p\" (UniqueName: \"kubernetes.io/projected/fef53718-de39-49b8-a9af-ef096a5e88ed-kube-api-access-tpb7p\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.460502 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpcsp\" (UniqueName: \"kubernetes.io/projected/295208bb-eebe-474d-9210-be3a67db28b0-kube-api-access-qpcsp\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.574998 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-f9d2p"] Sep 30 08:21:01 crc kubenswrapper[4810]: E0930 08:21:01.575596 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295208bb-eebe-474d-9210-be3a67db28b0" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.575610 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="295208bb-eebe-474d-9210-be3a67db28b0" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: E0930 08:21:01.575634 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef53718-de39-49b8-a9af-ef096a5e88ed" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.575640 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef53718-de39-49b8-a9af-ef096a5e88ed" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: E0930 08:21:01.575659 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af23fdb8-4a62-427f-8197-ca7a551b7b22" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.575666 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="af23fdb8-4a62-427f-8197-ca7a551b7b22" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.576114 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fef53718-de39-49b8-a9af-ef096a5e88ed" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.576132 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="295208bb-eebe-474d-9210-be3a67db28b0" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.576166 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="af23fdb8-4a62-427f-8197-ca7a551b7b22" containerName="mariadb-database-create" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.578101 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f9d2p" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.587781 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-f9d2p"] Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.635985 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c7s2d" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.664994 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hn9m\" (UniqueName: \"kubernetes.io/projected/0834a080-3a6e-43c5-95d9-c60c4a10aa66-kube-api-access-5hn9m\") pod \"placement-db-create-f9d2p\" (UID: \"0834a080-3a6e-43c5-95d9-c60c4a10aa66\") " pod="openstack/placement-db-create-f9d2p" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.692410 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-95ds9" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.693119 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-95ds9" event={"ID":"af23fdb8-4a62-427f-8197-ca7a551b7b22","Type":"ContainerDied","Data":"c1c94ed971921a329482c622028d3dbee0f5a5e437cd37435737b834920f7303"} Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.693149 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1c94ed971921a329482c622028d3dbee0f5a5e437cd37435737b834920f7303" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.699131 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jgr2f" event={"ID":"fef53718-de39-49b8-a9af-ef096a5e88ed","Type":"ContainerDied","Data":"47438bbd3f5226677ff0404ee936923f60f768df8ae733480702da235b9cc857"} Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.699250 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47438bbd3f5226677ff0404ee936923f60f768df8ae733480702da235b9cc857" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.699432 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jgr2f" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.713533 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"2acde81cff846b04e6548cf31bca70941f79857376cfd42edd9b39e7484f71b3"} Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.713575 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"07bc5b90640dc457ec2b4f486048871321556e8340afc557e126186b0fefc88c"} Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.713584 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"557a63c66c39af8869c1bd3494152344f496bff8dcf9d79261b4f4c3c5928cfe"} Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.715341 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tdcfg" event={"ID":"295208bb-eebe-474d-9210-be3a67db28b0","Type":"ContainerDied","Data":"3450f901264838c3bfbba200f42c0cb0e8daa70d7c985642ac5e4a6bad564adf"} Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.715361 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3450f901264838c3bfbba200f42c0cb0e8daa70d7c985642ac5e4a6bad564adf" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.717056 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tdcfg" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.755988 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.758570 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="prometheus" containerID="cri-o://41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e" gracePeriod=600 Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.758834 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="thanos-sidecar" containerID="cri-o://003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a" gracePeriod=600 Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.758941 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="config-reloader" containerID="cri-o://93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703" gracePeriod=600 Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.766326 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hn9m\" (UniqueName: \"kubernetes.io/projected/0834a080-3a6e-43c5-95d9-c60c4a10aa66-kube-api-access-5hn9m\") pod \"placement-db-create-f9d2p\" (UID: \"0834a080-3a6e-43c5-95d9-c60c4a10aa66\") " pod="openstack/placement-db-create-f9d2p" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.838194 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hn9m\" (UniqueName: \"kubernetes.io/projected/0834a080-3a6e-43c5-95d9-c60c4a10aa66-kube-api-access-5hn9m\") pod \"placement-db-create-f9d2p\" (UID: \"0834a080-3a6e-43c5-95d9-c60c4a10aa66\") " pod="openstack/placement-db-create-f9d2p" Sep 30 08:21:01 crc kubenswrapper[4810]: I0930 08:21:01.962713 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f9d2p" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.232958 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.286712 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run\") pod \"12c190a8-686e-4984-886e-ab7fb72f3d0e\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.286854 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run" (OuterVolumeSpecName: "var-run") pod "12c190a8-686e-4984-886e-ab7fb72f3d0e" (UID: "12c190a8-686e-4984-886e-ab7fb72f3d0e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.286896 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zblvq\" (UniqueName: \"kubernetes.io/projected/12c190a8-686e-4984-886e-ab7fb72f3d0e-kube-api-access-zblvq\") pod \"12c190a8-686e-4984-886e-ab7fb72f3d0e\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.286918 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-log-ovn\") pod \"12c190a8-686e-4984-886e-ab7fb72f3d0e\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.286965 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run-ovn\") pod \"12c190a8-686e-4984-886e-ab7fb72f3d0e\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.287001 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-scripts\") pod \"12c190a8-686e-4984-886e-ab7fb72f3d0e\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.287036 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-additional-scripts\") pod \"12c190a8-686e-4984-886e-ab7fb72f3d0e\" (UID: \"12c190a8-686e-4984-886e-ab7fb72f3d0e\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.287376 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.287499 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "12c190a8-686e-4984-886e-ab7fb72f3d0e" (UID: "12c190a8-686e-4984-886e-ab7fb72f3d0e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.287549 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "12c190a8-686e-4984-886e-ab7fb72f3d0e" (UID: "12c190a8-686e-4984-886e-ab7fb72f3d0e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.287931 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "12c190a8-686e-4984-886e-ab7fb72f3d0e" (UID: "12c190a8-686e-4984-886e-ab7fb72f3d0e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.288250 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-scripts" (OuterVolumeSpecName: "scripts") pod "12c190a8-686e-4984-886e-ab7fb72f3d0e" (UID: "12c190a8-686e-4984-886e-ab7fb72f3d0e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.305463 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c190a8-686e-4984-886e-ab7fb72f3d0e-kube-api-access-zblvq" (OuterVolumeSpecName: "kube-api-access-zblvq") pod "12c190a8-686e-4984-886e-ab7fb72f3d0e" (UID: "12c190a8-686e-4984-886e-ab7fb72f3d0e"). InnerVolumeSpecName "kube-api-access-zblvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.388483 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zblvq\" (UniqueName: \"kubernetes.io/projected/12c190a8-686e-4984-886e-ab7fb72f3d0e-kube-api-access-zblvq\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.388522 4810 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.388533 4810 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/12c190a8-686e-4984-886e-ab7fb72f3d0e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.388541 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.388550 4810 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/12c190a8-686e-4984-886e-ab7fb72f3d0e-additional-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.431213 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-c7s2d"] Sep 30 08:21:02 crc kubenswrapper[4810]: W0930 08:21:02.455565 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7af7ec1f_cde8_4a71_953d_ed7426ba7727.slice/crio-8b708a0b6b6bb3903309a4f540ef869af52e42b3067c6ae17d79c1598d6c87bb WatchSource:0}: Error finding container 8b708a0b6b6bb3903309a4f540ef869af52e42b3067c6ae17d79c1598d6c87bb: Status 404 returned error can't find the container with id 8b708a0b6b6bb3903309a4f540ef869af52e42b3067c6ae17d79c1598d6c87bb Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.490909 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-9q2d8" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.620081 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-f9d2p"] Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.622695 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.692464 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e9cd2658-4276-4957-9941-ab397d3e0d5f-config-out\") pod \"e9cd2658-4276-4957-9941-ab397d3e0d5f\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.692604 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"e9cd2658-4276-4957-9941-ab397d3e0d5f\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.692627 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-thanos-prometheus-http-client-file\") pod \"e9cd2658-4276-4957-9941-ab397d3e0d5f\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.692674 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-tls-assets\") pod \"e9cd2658-4276-4957-9941-ab397d3e0d5f\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.692736 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-web-config\") pod \"e9cd2658-4276-4957-9941-ab397d3e0d5f\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.692765 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-config\") pod \"e9cd2658-4276-4957-9941-ab397d3e0d5f\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.692804 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e9cd2658-4276-4957-9941-ab397d3e0d5f-prometheus-metric-storage-rulefiles-0\") pod \"e9cd2658-4276-4957-9941-ab397d3e0d5f\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.692823 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw8c9\" (UniqueName: \"kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-kube-api-access-vw8c9\") pod \"e9cd2658-4276-4957-9941-ab397d3e0d5f\" (UID: \"e9cd2658-4276-4957-9941-ab397d3e0d5f\") " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.696440 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9cd2658-4276-4957-9941-ab397d3e0d5f-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "e9cd2658-4276-4957-9941-ab397d3e0d5f" (UID: "e9cd2658-4276-4957-9941-ab397d3e0d5f"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.702236 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "e9cd2658-4276-4957-9941-ab397d3e0d5f" (UID: "e9cd2658-4276-4957-9941-ab397d3e0d5f"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.702295 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9cd2658-4276-4957-9941-ab397d3e0d5f-config-out" (OuterVolumeSpecName: "config-out") pod "e9cd2658-4276-4957-9941-ab397d3e0d5f" (UID: "e9cd2658-4276-4957-9941-ab397d3e0d5f"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.703707 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-kube-api-access-vw8c9" (OuterVolumeSpecName: "kube-api-access-vw8c9") pod "e9cd2658-4276-4957-9941-ab397d3e0d5f" (UID: "e9cd2658-4276-4957-9941-ab397d3e0d5f"). InnerVolumeSpecName "kube-api-access-vw8c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.711897 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-config" (OuterVolumeSpecName: "config") pod "e9cd2658-4276-4957-9941-ab397d3e0d5f" (UID: "e9cd2658-4276-4957-9941-ab397d3e0d5f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.712205 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "e9cd2658-4276-4957-9941-ab397d3e0d5f" (UID: "e9cd2658-4276-4957-9941-ab397d3e0d5f"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.741079 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "e9cd2658-4276-4957-9941-ab397d3e0d5f" (UID: "e9cd2658-4276-4957-9941-ab397d3e0d5f"). InnerVolumeSpecName "pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.742633 4810 generic.go:334] "Generic (PLEG): container finished" podID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerID="003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a" exitCode=0 Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.742673 4810 generic.go:334] "Generic (PLEG): container finished" podID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerID="93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703" exitCode=0 Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.742709 4810 generic.go:334] "Generic (PLEG): container finished" podID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerID="41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e" exitCode=0 Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.742840 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.742905 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e9cd2658-4276-4957-9941-ab397d3e0d5f","Type":"ContainerDied","Data":"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a"} Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.742939 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e9cd2658-4276-4957-9941-ab397d3e0d5f","Type":"ContainerDied","Data":"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703"} Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.742951 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e9cd2658-4276-4957-9941-ab397d3e0d5f","Type":"ContainerDied","Data":"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e"} Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.742963 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e9cd2658-4276-4957-9941-ab397d3e0d5f","Type":"ContainerDied","Data":"fd53bdebccb0da1f617276a46b96b776d9408f63afda4bbf1230aaa839f976d0"} Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.742979 4810 scope.go:117] "RemoveContainer" containerID="003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.746504 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-web-config" (OuterVolumeSpecName: "web-config") pod "e9cd2658-4276-4957-9941-ab397d3e0d5f" (UID: "e9cd2658-4276-4957-9941-ab397d3e0d5f"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.750060 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"2cf14ace5ac289d4a07345b83c086519873f778c770160f5e5a24adfd9decbb8"} Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.751441 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-f9d2p" event={"ID":"0834a080-3a6e-43c5-95d9-c60c4a10aa66","Type":"ContainerStarted","Data":"9e286f2c278346d905ae04a9d7b9a9990684d4b7f1f7830a4ca0213e3f0e4922"} Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.753419 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8-config-x5ckn" event={"ID":"12c190a8-686e-4984-886e-ab7fb72f3d0e","Type":"ContainerDied","Data":"4eb6fd3fce833bbde787e0a56aa307dbc5c513848400c0d8467bdc1a92e7ee8d"} Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.753456 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eb6fd3fce833bbde787e0a56aa307dbc5c513848400c0d8467bdc1a92e7ee8d" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.753510 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-x5ckn" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.769072 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-c7s2d" event={"ID":"7af7ec1f-cde8-4a71-953d-ed7426ba7727","Type":"ContainerStarted","Data":"65ba002271130d507ec502565bd2b4c5eb79b9fc83389096ba418751bce7be6a"} Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.770820 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-c7s2d" event={"ID":"7af7ec1f-cde8-4a71-953d-ed7426ba7727","Type":"ContainerStarted","Data":"8b708a0b6b6bb3903309a4f540ef869af52e42b3067c6ae17d79c1598d6c87bb"} Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.796395 4810 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e9cd2658-4276-4957-9941-ab397d3e0d5f-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.796436 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw8c9\" (UniqueName: \"kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-kube-api-access-vw8c9\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.796453 4810 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e9cd2658-4276-4957-9941-ab397d3e0d5f-config-out\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.796484 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") on node \"crc\" " Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.796499 4810 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.796511 4810 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e9cd2658-4276-4957-9941-ab397d3e0d5f-tls-assets\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.796527 4810 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-web-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.796539 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e9cd2658-4276-4957-9941-ab397d3e0d5f-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.798989 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-9q2d8-config-x5ckn"] Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.810985 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-9q2d8-config-x5ckn"] Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.819141 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-c7s2d" podStartSLOduration=1.8191213990000001 podStartE2EDuration="1.819121399s" podCreationTimestamp="2025-09-30 08:21:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:21:02.78889585 +0000 UTC m=+1086.241095117" watchObservedRunningTime="2025-09-30 08:21:02.819121399 +0000 UTC m=+1086.271320666" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.835911 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9q2d8-config-ftb9s"] Sep 30 08:21:02 crc kubenswrapper[4810]: E0930 08:21:02.836301 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="thanos-sidecar" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.836313 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="thanos-sidecar" Sep 30 08:21:02 crc kubenswrapper[4810]: E0930 08:21:02.836328 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="init-config-reloader" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.836336 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="init-config-reloader" Sep 30 08:21:02 crc kubenswrapper[4810]: E0930 08:21:02.836349 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="prometheus" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.836356 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="prometheus" Sep 30 08:21:02 crc kubenswrapper[4810]: E0930 08:21:02.836373 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="config-reloader" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.836379 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="config-reloader" Sep 30 08:21:02 crc kubenswrapper[4810]: E0930 08:21:02.836388 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c190a8-686e-4984-886e-ab7fb72f3d0e" containerName="ovn-config" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.836394 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c190a8-686e-4984-886e-ab7fb72f3d0e" containerName="ovn-config" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.836560 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="prometheus" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.836574 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="config-reloader" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.836589 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c190a8-686e-4984-886e-ab7fb72f3d0e" containerName="ovn-config" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.836605 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" containerName="thanos-sidecar" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.837129 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.839420 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.840085 4810 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.840238 4810 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c") on node "crc" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.855002 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9q2d8-config-ftb9s"] Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.897609 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run-ovn\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.897671 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2jwj\" (UniqueName: \"kubernetes.io/projected/1803bddd-901e-48bd-a1e9-50a3132e62e3-kube-api-access-h2jwj\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.897694 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.897928 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-scripts\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.897989 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-log-ovn\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.898073 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-additional-scripts\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.898241 4810 reconciler_common.go:293] "Volume detached for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.999603 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2jwj\" (UniqueName: \"kubernetes.io/projected/1803bddd-901e-48bd-a1e9-50a3132e62e3-kube-api-access-h2jwj\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.999656 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:02 crc kubenswrapper[4810]: I0930 08:21:02.999745 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-scripts\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:02.999811 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-log-ovn\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:02.999840 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-additional-scripts\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:02.999876 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run-ovn\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.000254 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-log-ovn\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.000297 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.001054 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-additional-scripts\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.001131 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run-ovn\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.002141 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-scripts\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.020150 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2jwj\" (UniqueName: \"kubernetes.io/projected/1803bddd-901e-48bd-a1e9-50a3132e62e3-kube-api-access-h2jwj\") pod \"ovn-controller-9q2d8-config-ftb9s\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.085651 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.110165 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.115856 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.118245 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.125423 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.128334 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.139387 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-t6zzk" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.139654 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.139816 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.140584 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.140961 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.141148 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.218167 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.250153 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-b9dc-account-create-sxxkk"] Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.251820 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-b9dc-account-create-sxxkk" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.254509 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.257544 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-b9dc-account-create-sxxkk"] Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.306051 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.306107 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.306135 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.306172 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.306214 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eb9bba80-4046-4788-ac20-c682625132d8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.306236 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.306296 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.306321 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.306346 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24d77\" (UniqueName: \"kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-kube-api-access-24d77\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.306415 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.306442 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eb9bba80-4046-4788-ac20-c682625132d8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.309212 4810 scope.go:117] "RemoveContainer" containerID="93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.316074 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c190a8-686e-4984-886e-ab7fb72f3d0e" path="/var/lib/kubelet/pods/12c190a8-686e-4984-886e-ab7fb72f3d0e/volumes" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.316792 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9cd2658-4276-4957-9941-ab397d3e0d5f" path="/var/lib/kubelet/pods/e9cd2658-4276-4957-9941-ab397d3e0d5f/volumes" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.369635 4810 scope.go:117] "RemoveContainer" containerID="41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408021 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwprn\" (UniqueName: \"kubernetes.io/projected/7ab68570-8ed3-4c73-94da-88ced0b8d75e-kube-api-access-nwprn\") pod \"watcher-b9dc-account-create-sxxkk\" (UID: \"7ab68570-8ed3-4c73-94da-88ced0b8d75e\") " pod="openstack/watcher-b9dc-account-create-sxxkk" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408086 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408111 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eb9bba80-4046-4788-ac20-c682625132d8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408160 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408181 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408207 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408235 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408297 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eb9bba80-4046-4788-ac20-c682625132d8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408328 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408401 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408429 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.408485 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24d77\" (UniqueName: \"kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-kube-api-access-24d77\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.409458 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eb9bba80-4046-4788-ac20-c682625132d8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.416386 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.416436 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b8ca484aa6b3c023d84489054d30b21e25eca5ebf5935b3c554e8125ba6153af/globalmount\"" pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.418604 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.418847 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eb9bba80-4046-4788-ac20-c682625132d8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.419220 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.420875 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.443204 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.443214 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.443816 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.453620 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24d77\" (UniqueName: \"kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-kube-api-access-24d77\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.453845 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.486208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.492208 4810 scope.go:117] "RemoveContainer" containerID="e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.510564 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwprn\" (UniqueName: \"kubernetes.io/projected/7ab68570-8ed3-4c73-94da-88ced0b8d75e-kube-api-access-nwprn\") pod \"watcher-b9dc-account-create-sxxkk\" (UID: \"7ab68570-8ed3-4c73-94da-88ced0b8d75e\") " pod="openstack/watcher-b9dc-account-create-sxxkk" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.535641 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwprn\" (UniqueName: \"kubernetes.io/projected/7ab68570-8ed3-4c73-94da-88ced0b8d75e-kube-api-access-nwprn\") pod \"watcher-b9dc-account-create-sxxkk\" (UID: \"7ab68570-8ed3-4c73-94da-88ced0b8d75e\") " pod="openstack/watcher-b9dc-account-create-sxxkk" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.548038 4810 scope.go:117] "RemoveContainer" containerID="003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a" Sep 30 08:21:03 crc kubenswrapper[4810]: E0930 08:21:03.548786 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a\": container with ID starting with 003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a not found: ID does not exist" containerID="003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.548840 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a"} err="failed to get container status \"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a\": rpc error: code = NotFound desc = could not find container \"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a\": container with ID starting with 003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.548869 4810 scope.go:117] "RemoveContainer" containerID="93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703" Sep 30 08:21:03 crc kubenswrapper[4810]: E0930 08:21:03.549191 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703\": container with ID starting with 93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703 not found: ID does not exist" containerID="93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.549213 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703"} err="failed to get container status \"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703\": rpc error: code = NotFound desc = could not find container \"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703\": container with ID starting with 93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703 not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.549252 4810 scope.go:117] "RemoveContainer" containerID="41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e" Sep 30 08:21:03 crc kubenswrapper[4810]: E0930 08:21:03.549561 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e\": container with ID starting with 41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e not found: ID does not exist" containerID="41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.549581 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e"} err="failed to get container status \"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e\": rpc error: code = NotFound desc = could not find container \"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e\": container with ID starting with 41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.549617 4810 scope.go:117] "RemoveContainer" containerID="e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8" Sep 30 08:21:03 crc kubenswrapper[4810]: E0930 08:21:03.551400 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8\": container with ID starting with e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8 not found: ID does not exist" containerID="e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.551450 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8"} err="failed to get container status \"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8\": rpc error: code = NotFound desc = could not find container \"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8\": container with ID starting with e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8 not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.551470 4810 scope.go:117] "RemoveContainer" containerID="003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.555064 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a"} err="failed to get container status \"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a\": rpc error: code = NotFound desc = could not find container \"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a\": container with ID starting with 003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.555097 4810 scope.go:117] "RemoveContainer" containerID="93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.559358 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703"} err="failed to get container status \"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703\": rpc error: code = NotFound desc = could not find container \"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703\": container with ID starting with 93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703 not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.559396 4810 scope.go:117] "RemoveContainer" containerID="41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.559725 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e"} err="failed to get container status \"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e\": rpc error: code = NotFound desc = could not find container \"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e\": container with ID starting with 41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.559747 4810 scope.go:117] "RemoveContainer" containerID="e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.560471 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8"} err="failed to get container status \"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8\": rpc error: code = NotFound desc = could not find container \"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8\": container with ID starting with e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8 not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.560494 4810 scope.go:117] "RemoveContainer" containerID="003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.560876 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a"} err="failed to get container status \"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a\": rpc error: code = NotFound desc = could not find container \"003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a\": container with ID starting with 003c05888819201881526e22d0463a0de978d90a7350d8afad698b25c3a42e6a not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.560907 4810 scope.go:117] "RemoveContainer" containerID="93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.561236 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703"} err="failed to get container status \"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703\": rpc error: code = NotFound desc = could not find container \"93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703\": container with ID starting with 93489d92758b11781f9f8336f0a477c7da3cd6744d071ed5c3f1d130c78fa703 not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.561287 4810 scope.go:117] "RemoveContainer" containerID="41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.561571 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e"} err="failed to get container status \"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e\": rpc error: code = NotFound desc = could not find container \"41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e\": container with ID starting with 41269f707214bd9902ac9ba14601d3668bea7fdcc617066d406925d7ea48819e not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.561594 4810 scope.go:117] "RemoveContainer" containerID="e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.561938 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8"} err="failed to get container status \"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8\": rpc error: code = NotFound desc = could not find container \"e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8\": container with ID starting with e6691f21bd0697df9cdb9ad2ed77408344818f96389c09d00d19ea7b293d15b8 not found: ID does not exist" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.574470 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-b9dc-account-create-sxxkk" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.753175 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.781912 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"522d533a38d15c4339a83cf81590eb4d41e1f28203f9aeef884f012e2a24b496"} Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.781965 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"e8cf0b0e0e62c71b370366ef834221e73980f233b527687ba8932bb6770aa92a"} Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.783593 4810 generic.go:334] "Generic (PLEG): container finished" podID="0834a080-3a6e-43c5-95d9-c60c4a10aa66" containerID="5fa449acb9ced2efab045b346eb0ccd4d51e48bd2b87ceab7395768c4f9460dd" exitCode=0 Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.783659 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-f9d2p" event={"ID":"0834a080-3a6e-43c5-95d9-c60c4a10aa66","Type":"ContainerDied","Data":"5fa449acb9ced2efab045b346eb0ccd4d51e48bd2b87ceab7395768c4f9460dd"} Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.786599 4810 generic.go:334] "Generic (PLEG): container finished" podID="7af7ec1f-cde8-4a71-953d-ed7426ba7727" containerID="65ba002271130d507ec502565bd2b4c5eb79b9fc83389096ba418751bce7be6a" exitCode=0 Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.786687 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-c7s2d" event={"ID":"7af7ec1f-cde8-4a71-953d-ed7426ba7727","Type":"ContainerDied","Data":"65ba002271130d507ec502565bd2b4c5eb79b9fc83389096ba418751bce7be6a"} Sep 30 08:21:03 crc kubenswrapper[4810]: I0930 08:21:03.832833 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9q2d8-config-ftb9s"] Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.019622 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-b9dc-account-create-sxxkk"] Sep 30 08:21:04 crc kubenswrapper[4810]: W0930 08:21:04.045906 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ab68570_8ed3_4c73_94da_88ced0b8d75e.slice/crio-b9cb0d9c9a0d7af8859db2eaaef0f6946add69ddc114a831ef85426a3e52193b WatchSource:0}: Error finding container b9cb0d9c9a0d7af8859db2eaaef0f6946add69ddc114a831ef85426a3e52193b: Status 404 returned error can't find the container with id b9cb0d9c9a0d7af8859db2eaaef0f6946add69ddc114a831ef85426a3e52193b Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.247642 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:21:04 crc kubenswrapper[4810]: W0930 08:21:04.251112 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb9bba80_4046_4788_ac20_c682625132d8.slice/crio-adb297544aef1eeff8c20225e20a3e7b5541f56f960469d3e41d23e08b6ebc18 WatchSource:0}: Error finding container adb297544aef1eeff8c20225e20a3e7b5541f56f960469d3e41d23e08b6ebc18: Status 404 returned error can't find the container with id adb297544aef1eeff8c20225e20a3e7b5541f56f960469d3e41d23e08b6ebc18 Sep 30 08:21:04 crc kubenswrapper[4810]: E0930 08:21:04.753332 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dacf67c_34b5_4c58_9f8d_00adfdcf74b7.slice\": RecentStats: unable to find data in memory cache]" Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.806048 4810 generic.go:334] "Generic (PLEG): container finished" podID="7ab68570-8ed3-4c73-94da-88ced0b8d75e" containerID="3c0e2d67e5392c64e8647ef5661596e0c743710de0f86368f60b95241e617e0a" exitCode=0 Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.806334 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-b9dc-account-create-sxxkk" event={"ID":"7ab68570-8ed3-4c73-94da-88ced0b8d75e","Type":"ContainerDied","Data":"3c0e2d67e5392c64e8647ef5661596e0c743710de0f86368f60b95241e617e0a"} Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.806466 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-b9dc-account-create-sxxkk" event={"ID":"7ab68570-8ed3-4c73-94da-88ced0b8d75e","Type":"ContainerStarted","Data":"b9cb0d9c9a0d7af8859db2eaaef0f6946add69ddc114a831ef85426a3e52193b"} Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.827131 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"e3f0637499880ec5ca80a9853fee086efa17c0e55a3915cae9d505b39a13e67e"} Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.827184 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"3a9856267724c3116e1f77d6ec25e2ea8efafad1b27992f1c9e12a897d7876cc"} Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.827196 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"ebd5e91202884dc819cc6ebe2a83affec8df98fb7bc208cff15defe2c07afce8"} Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.827209 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"f78a9e3ca64b3664bce92a85a853a9c60b96fc16db24ba72ce36fc1f49f07dab"} Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.828879 4810 generic.go:334] "Generic (PLEG): container finished" podID="1803bddd-901e-48bd-a1e9-50a3132e62e3" containerID="9701f0be78a5ff9e61d96211139edfd486723cff8afe5a4c456f626be4594931" exitCode=0 Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.828932 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8-config-ftb9s" event={"ID":"1803bddd-901e-48bd-a1e9-50a3132e62e3","Type":"ContainerDied","Data":"9701f0be78a5ff9e61d96211139edfd486723cff8afe5a4c456f626be4594931"} Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.828953 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8-config-ftb9s" event={"ID":"1803bddd-901e-48bd-a1e9-50a3132e62e3","Type":"ContainerStarted","Data":"f75a391584d3316056ada0977f500ec04957f8ddc3baae8904b949e279215153"} Sep 30 08:21:04 crc kubenswrapper[4810]: I0930 08:21:04.830599 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb9bba80-4046-4788-ac20-c682625132d8","Type":"ContainerStarted","Data":"adb297544aef1eeff8c20225e20a3e7b5541f56f960469d3e41d23e08b6ebc18"} Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.307786 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f9d2p" Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.315301 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c7s2d" Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.444685 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hn9m\" (UniqueName: \"kubernetes.io/projected/0834a080-3a6e-43c5-95d9-c60c4a10aa66-kube-api-access-5hn9m\") pod \"0834a080-3a6e-43c5-95d9-c60c4a10aa66\" (UID: \"0834a080-3a6e-43c5-95d9-c60c4a10aa66\") " Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.444815 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5gj2\" (UniqueName: \"kubernetes.io/projected/7af7ec1f-cde8-4a71-953d-ed7426ba7727-kube-api-access-h5gj2\") pod \"7af7ec1f-cde8-4a71-953d-ed7426ba7727\" (UID: \"7af7ec1f-cde8-4a71-953d-ed7426ba7727\") " Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.453916 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7af7ec1f-cde8-4a71-953d-ed7426ba7727-kube-api-access-h5gj2" (OuterVolumeSpecName: "kube-api-access-h5gj2") pod "7af7ec1f-cde8-4a71-953d-ed7426ba7727" (UID: "7af7ec1f-cde8-4a71-953d-ed7426ba7727"). InnerVolumeSpecName "kube-api-access-h5gj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.454023 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0834a080-3a6e-43c5-95d9-c60c4a10aa66-kube-api-access-5hn9m" (OuterVolumeSpecName: "kube-api-access-5hn9m") pod "0834a080-3a6e-43c5-95d9-c60c4a10aa66" (UID: "0834a080-3a6e-43c5-95d9-c60c4a10aa66"). InnerVolumeSpecName "kube-api-access-5hn9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.546706 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hn9m\" (UniqueName: \"kubernetes.io/projected/0834a080-3a6e-43c5-95d9-c60c4a10aa66-kube-api-access-5hn9m\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.546742 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5gj2\" (UniqueName: \"kubernetes.io/projected/7af7ec1f-cde8-4a71-953d-ed7426ba7727-kube-api-access-h5gj2\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.843221 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f9d2p" Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.843322 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-f9d2p" event={"ID":"0834a080-3a6e-43c5-95d9-c60c4a10aa66","Type":"ContainerDied","Data":"9e286f2c278346d905ae04a9d7b9a9990684d4b7f1f7830a4ca0213e3f0e4922"} Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.843372 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e286f2c278346d905ae04a9d7b9a9990684d4b7f1f7830a4ca0213e3f0e4922" Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.853800 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-c7s2d" event={"ID":"7af7ec1f-cde8-4a71-953d-ed7426ba7727","Type":"ContainerDied","Data":"8b708a0b6b6bb3903309a4f540ef869af52e42b3067c6ae17d79c1598d6c87bb"} Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.853849 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b708a0b6b6bb3903309a4f540ef869af52e42b3067c6ae17d79c1598d6c87bb" Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.853847 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c7s2d" Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.867079 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edd67220-e97a-465e-9671-2cfcf760475a","Type":"ContainerStarted","Data":"e1ed00edcfaa950fcf8d76632c7117bbd614ddfc76e309b43b1debc5796693ee"} Sep 30 08:21:05 crc kubenswrapper[4810]: I0930 08:21:05.907433 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.687034781 podStartE2EDuration="42.907413715s" podCreationTimestamp="2025-09-30 08:20:23 +0000 UTC" firstStartedPulling="2025-09-30 08:20:57.16062527 +0000 UTC m=+1080.612824537" lastFinishedPulling="2025-09-30 08:21:03.381004204 +0000 UTC m=+1086.833203471" observedRunningTime="2025-09-30 08:21:05.905583591 +0000 UTC m=+1089.357782858" watchObservedRunningTime="2025-09-30 08:21:05.907413715 +0000 UTC m=+1089.359613002" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.224532 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b9b87fb55-7jxzz"] Sep 30 08:21:06 crc kubenswrapper[4810]: E0930 08:21:06.224906 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af7ec1f-cde8-4a71-953d-ed7426ba7727" containerName="mariadb-database-create" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.224922 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af7ec1f-cde8-4a71-953d-ed7426ba7727" containerName="mariadb-database-create" Sep 30 08:21:06 crc kubenswrapper[4810]: E0930 08:21:06.224941 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0834a080-3a6e-43c5-95d9-c60c4a10aa66" containerName="mariadb-database-create" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.224947 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0834a080-3a6e-43c5-95d9-c60c4a10aa66" containerName="mariadb-database-create" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.225151 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7af7ec1f-cde8-4a71-953d-ed7426ba7727" containerName="mariadb-database-create" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.225168 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0834a080-3a6e-43c5-95d9-c60c4a10aa66" containerName="mariadb-database-create" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.226045 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.229829 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.240248 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b9b87fb55-7jxzz"] Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.243184 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.359649 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-log-ovn\") pod \"1803bddd-901e-48bd-a1e9-50a3132e62e3\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.359728 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1803bddd-901e-48bd-a1e9-50a3132e62e3" (UID: "1803bddd-901e-48bd-a1e9-50a3132e62e3"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.359789 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-additional-scripts\") pod \"1803bddd-901e-48bd-a1e9-50a3132e62e3\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.359838 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2jwj\" (UniqueName: \"kubernetes.io/projected/1803bddd-901e-48bd-a1e9-50a3132e62e3-kube-api-access-h2jwj\") pod \"1803bddd-901e-48bd-a1e9-50a3132e62e3\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.359872 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-scripts\") pod \"1803bddd-901e-48bd-a1e9-50a3132e62e3\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.359931 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run-ovn\") pod \"1803bddd-901e-48bd-a1e9-50a3132e62e3\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360072 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1803bddd-901e-48bd-a1e9-50a3132e62e3" (UID: "1803bddd-901e-48bd-a1e9-50a3132e62e3"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360151 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run\") pod \"1803bddd-901e-48bd-a1e9-50a3132e62e3\" (UID: \"1803bddd-901e-48bd-a1e9-50a3132e62e3\") " Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360201 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run" (OuterVolumeSpecName: "var-run") pod "1803bddd-901e-48bd-a1e9-50a3132e62e3" (UID: "1803bddd-901e-48bd-a1e9-50a3132e62e3"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360546 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbd8v\" (UniqueName: \"kubernetes.io/projected/cf25d291-3d0d-48c0-9dae-405ef9fd4922-kube-api-access-cbd8v\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360575 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-svc\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360581 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "1803bddd-901e-48bd-a1e9-50a3132e62e3" (UID: "1803bddd-901e-48bd-a1e9-50a3132e62e3"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360679 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-swift-storage-0\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360741 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-sb\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360759 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-config\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360838 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-nb\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360903 4810 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-additional-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360910 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-scripts" (OuterVolumeSpecName: "scripts") pod "1803bddd-901e-48bd-a1e9-50a3132e62e3" (UID: "1803bddd-901e-48bd-a1e9-50a3132e62e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.360932 4810 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.361004 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-run\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.361020 4810 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1803bddd-901e-48bd-a1e9-50a3132e62e3-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.368008 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1803bddd-901e-48bd-a1e9-50a3132e62e3-kube-api-access-h2jwj" (OuterVolumeSpecName: "kube-api-access-h2jwj") pod "1803bddd-901e-48bd-a1e9-50a3132e62e3" (UID: "1803bddd-901e-48bd-a1e9-50a3132e62e3"). InnerVolumeSpecName "kube-api-access-h2jwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.388490 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-b9dc-account-create-sxxkk" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.461958 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-nb\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.462097 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbd8v\" (UniqueName: \"kubernetes.io/projected/cf25d291-3d0d-48c0-9dae-405ef9fd4922-kube-api-access-cbd8v\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.462122 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-svc\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.462171 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-swift-storage-0\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.462192 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-sb\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.462213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-config\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.462278 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2jwj\" (UniqueName: \"kubernetes.io/projected/1803bddd-901e-48bd-a1e9-50a3132e62e3-kube-api-access-h2jwj\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.462289 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1803bddd-901e-48bd-a1e9-50a3132e62e3-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.465296 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-sb\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.465417 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-svc\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.466516 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-nb\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.467002 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-swift-storage-0\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.467642 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-config\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.488531 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbd8v\" (UniqueName: \"kubernetes.io/projected/cf25d291-3d0d-48c0-9dae-405ef9fd4922-kube-api-access-cbd8v\") pod \"dnsmasq-dns-5b9b87fb55-7jxzz\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.561859 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.562674 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwprn\" (UniqueName: \"kubernetes.io/projected/7ab68570-8ed3-4c73-94da-88ced0b8d75e-kube-api-access-nwprn\") pod \"7ab68570-8ed3-4c73-94da-88ced0b8d75e\" (UID: \"7ab68570-8ed3-4c73-94da-88ced0b8d75e\") " Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.568147 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ab68570-8ed3-4c73-94da-88ced0b8d75e-kube-api-access-nwprn" (OuterVolumeSpecName: "kube-api-access-nwprn") pod "7ab68570-8ed3-4c73-94da-88ced0b8d75e" (UID: "7ab68570-8ed3-4c73-94da-88ced0b8d75e"). InnerVolumeSpecName "kube-api-access-nwprn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:06 crc kubenswrapper[4810]: I0930 08:21:06.664656 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwprn\" (UniqueName: \"kubernetes.io/projected/7ab68570-8ed3-4c73-94da-88ced0b8d75e-kube-api-access-nwprn\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.896491 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-ftb9s" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.899726 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8-config-ftb9s" event={"ID":"1803bddd-901e-48bd-a1e9-50a3132e62e3","Type":"ContainerDied","Data":"f75a391584d3316056ada0977f500ec04957f8ddc3baae8904b949e279215153"} Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.899784 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f75a391584d3316056ada0977f500ec04957f8ddc3baae8904b949e279215153" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.909253 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-588d-account-create-hl9zx"] Sep 30 08:21:07 crc kubenswrapper[4810]: E0930 08:21:06.909586 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ab68570-8ed3-4c73-94da-88ced0b8d75e" containerName="mariadb-account-create" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.909598 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ab68570-8ed3-4c73-94da-88ced0b8d75e" containerName="mariadb-account-create" Sep 30 08:21:07 crc kubenswrapper[4810]: E0930 08:21:06.909634 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1803bddd-901e-48bd-a1e9-50a3132e62e3" containerName="ovn-config" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.909640 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1803bddd-901e-48bd-a1e9-50a3132e62e3" containerName="ovn-config" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.909793 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1803bddd-901e-48bd-a1e9-50a3132e62e3" containerName="ovn-config" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.909806 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ab68570-8ed3-4c73-94da-88ced0b8d75e" containerName="mariadb-account-create" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.910334 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-588d-account-create-hl9zx" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.915174 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.920331 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-b9dc-account-create-sxxkk" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.923499 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-b9dc-account-create-sxxkk" event={"ID":"7ab68570-8ed3-4c73-94da-88ced0b8d75e","Type":"ContainerDied","Data":"b9cb0d9c9a0d7af8859db2eaaef0f6946add69ddc114a831ef85426a3e52193b"} Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.923533 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9cb0d9c9a0d7af8859db2eaaef0f6946add69ddc114a831ef85426a3e52193b" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.929730 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-588d-account-create-hl9zx"] Sep 30 08:21:07 crc kubenswrapper[4810]: W0930 08:21:06.948707 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf25d291_3d0d_48c0_9dae_405ef9fd4922.slice/crio-1ec309e8eace6dae459c65031b9fe45e3731b2545ed9529bbf89d0515acfa38b WatchSource:0}: Error finding container 1ec309e8eace6dae459c65031b9fe45e3731b2545ed9529bbf89d0515acfa38b: Status 404 returned error can't find the container with id 1ec309e8eace6dae459c65031b9fe45e3731b2545ed9529bbf89d0515acfa38b Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:06.951066 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b9b87fb55-7jxzz"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.070913 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmxwp\" (UniqueName: \"kubernetes.io/projected/127b94a4-4969-4a8d-b708-1b1c6f61d6af-kube-api-access-rmxwp\") pod \"glance-588d-account-create-hl9zx\" (UID: \"127b94a4-4969-4a8d-b708-1b1c6f61d6af\") " pod="openstack/glance-588d-account-create-hl9zx" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.174122 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmxwp\" (UniqueName: \"kubernetes.io/projected/127b94a4-4969-4a8d-b708-1b1c6f61d6af-kube-api-access-rmxwp\") pod \"glance-588d-account-create-hl9zx\" (UID: \"127b94a4-4969-4a8d-b708-1b1c6f61d6af\") " pod="openstack/glance-588d-account-create-hl9zx" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.197020 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmxwp\" (UniqueName: \"kubernetes.io/projected/127b94a4-4969-4a8d-b708-1b1c6f61d6af-kube-api-access-rmxwp\") pod \"glance-588d-account-create-hl9zx\" (UID: \"127b94a4-4969-4a8d-b708-1b1c6f61d6af\") " pod="openstack/glance-588d-account-create-hl9zx" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.241020 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-588d-account-create-hl9zx" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.339736 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-9q2d8-config-ftb9s"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.358786 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-9q2d8-config-ftb9s"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.448371 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9q2d8-config-dvzf6"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.452225 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.453262 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9q2d8-config-dvzf6"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.456129 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.586254 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-scripts\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.586326 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run-ovn\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.586379 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.586447 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-log-ovn\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.586468 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gjzw\" (UniqueName: \"kubernetes.io/projected/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-kube-api-access-8gjzw\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.586509 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-additional-scripts\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.646084 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-dfa1-account-create-tnx8l"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.647192 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dfa1-account-create-tnx8l" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.649778 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.656097 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-dfa1-account-create-tnx8l"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.687750 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-additional-scripts\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.687930 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-scripts\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.688003 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run-ovn\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.688061 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.688153 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-log-ovn\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.688188 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gjzw\" (UniqueName: \"kubernetes.io/projected/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-kube-api-access-8gjzw\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.688327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run-ovn\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.688355 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.688436 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-log-ovn\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.688717 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-additional-scripts\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.689979 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-scripts\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.705841 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gjzw\" (UniqueName: \"kubernetes.io/projected/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-kube-api-access-8gjzw\") pod \"ovn-controller-9q2d8-config-dvzf6\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.773440 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.789150 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7tkr\" (UniqueName: \"kubernetes.io/projected/d505f4a8-0fe5-4cc5-9e55-369f42edd90a-kube-api-access-b7tkr\") pod \"cinder-dfa1-account-create-tnx8l\" (UID: \"d505f4a8-0fe5-4cc5-9e55-369f42edd90a\") " pod="openstack/cinder-dfa1-account-create-tnx8l" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.842617 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-15e4-account-create-ds56p"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.843954 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-15e4-account-create-ds56p" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.848781 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.860388 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-15e4-account-create-ds56p"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.891609 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7tkr\" (UniqueName: \"kubernetes.io/projected/d505f4a8-0fe5-4cc5-9e55-369f42edd90a-kube-api-access-b7tkr\") pod \"cinder-dfa1-account-create-tnx8l\" (UID: \"d505f4a8-0fe5-4cc5-9e55-369f42edd90a\") " pod="openstack/cinder-dfa1-account-create-tnx8l" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.918376 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7tkr\" (UniqueName: \"kubernetes.io/projected/d505f4a8-0fe5-4cc5-9e55-369f42edd90a-kube-api-access-b7tkr\") pod \"cinder-dfa1-account-create-tnx8l\" (UID: \"d505f4a8-0fe5-4cc5-9e55-369f42edd90a\") " pod="openstack/cinder-dfa1-account-create-tnx8l" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.932025 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-588d-account-create-hl9zx"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.948457 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-410c-account-create-4czqf"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.953813 4810 generic.go:334] "Generic (PLEG): container finished" podID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerID="bfcc5073aa9d0282dace4806787c782cc58667fc7ab16add9a1d07edec3d8484" exitCode=0 Sep 30 08:21:07 crc kubenswrapper[4810]: W0930 08:21:07.954949 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod127b94a4_4969_4a8d_b708_1b1c6f61d6af.slice/crio-dd6fa11baee390f014380ff2b0ce56273fecbb95945ecd326f85e3197e36a0d1 WatchSource:0}: Error finding container dd6fa11baee390f014380ff2b0ce56273fecbb95945ecd326f85e3197e36a0d1: Status 404 returned error can't find the container with id dd6fa11baee390f014380ff2b0ce56273fecbb95945ecd326f85e3197e36a0d1 Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.955865 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb9bba80-4046-4788-ac20-c682625132d8","Type":"ContainerStarted","Data":"43a44738cda6d3139fcbce99404f6f250153b2205c24b1b2cda00b30fe0cb1ff"} Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.956232 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" event={"ID":"cf25d291-3d0d-48c0-9dae-405ef9fd4922","Type":"ContainerDied","Data":"bfcc5073aa9d0282dace4806787c782cc58667fc7ab16add9a1d07edec3d8484"} Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.956335 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" event={"ID":"cf25d291-3d0d-48c0-9dae-405ef9fd4922","Type":"ContainerStarted","Data":"1ec309e8eace6dae459c65031b9fe45e3731b2545ed9529bbf89d0515acfa38b"} Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.956522 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-410c-account-create-4czqf" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.958971 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.969607 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dfa1-account-create-tnx8l" Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.971505 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-410c-account-create-4czqf"] Sep 30 08:21:07 crc kubenswrapper[4810]: I0930 08:21:07.996399 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl7ct\" (UniqueName: \"kubernetes.io/projected/fcd581d7-6a83-4822-822b-421633417bdf-kube-api-access-bl7ct\") pod \"barbican-15e4-account-create-ds56p\" (UID: \"fcd581d7-6a83-4822-822b-421633417bdf\") " pod="openstack/barbican-15e4-account-create-ds56p" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.098241 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl7ct\" (UniqueName: \"kubernetes.io/projected/fcd581d7-6a83-4822-822b-421633417bdf-kube-api-access-bl7ct\") pod \"barbican-15e4-account-create-ds56p\" (UID: \"fcd581d7-6a83-4822-822b-421633417bdf\") " pod="openstack/barbican-15e4-account-create-ds56p" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.098313 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knpsj\" (UniqueName: \"kubernetes.io/projected/ca25948f-0522-43b1-9b6b-a7ce98207be0-kube-api-access-knpsj\") pod \"neutron-410c-account-create-4czqf\" (UID: \"ca25948f-0522-43b1-9b6b-a7ce98207be0\") " pod="openstack/neutron-410c-account-create-4czqf" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.117934 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl7ct\" (UniqueName: \"kubernetes.io/projected/fcd581d7-6a83-4822-822b-421633417bdf-kube-api-access-bl7ct\") pod \"barbican-15e4-account-create-ds56p\" (UID: \"fcd581d7-6a83-4822-822b-421633417bdf\") " pod="openstack/barbican-15e4-account-create-ds56p" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.198849 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-15e4-account-create-ds56p" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.199808 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knpsj\" (UniqueName: \"kubernetes.io/projected/ca25948f-0522-43b1-9b6b-a7ce98207be0-kube-api-access-knpsj\") pod \"neutron-410c-account-create-4czqf\" (UID: \"ca25948f-0522-43b1-9b6b-a7ce98207be0\") " pod="openstack/neutron-410c-account-create-4czqf" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.226091 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knpsj\" (UniqueName: \"kubernetes.io/projected/ca25948f-0522-43b1-9b6b-a7ce98207be0-kube-api-access-knpsj\") pod \"neutron-410c-account-create-4czqf\" (UID: \"ca25948f-0522-43b1-9b6b-a7ce98207be0\") " pod="openstack/neutron-410c-account-create-4czqf" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.289349 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9q2d8-config-dvzf6"] Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.499233 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-dfa1-account-create-tnx8l"] Sep 30 08:21:08 crc kubenswrapper[4810]: W0930 08:21:08.508422 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd505f4a8_0fe5_4cc5_9e55_369f42edd90a.slice/crio-c69c8501aaf38975f10e8cce7330a59c2a5049161f934259eec8414296e3530e WatchSource:0}: Error finding container c69c8501aaf38975f10e8cce7330a59c2a5049161f934259eec8414296e3530e: Status 404 returned error can't find the container with id c69c8501aaf38975f10e8cce7330a59c2a5049161f934259eec8414296e3530e Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.521581 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-410c-account-create-4czqf" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.603649 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-2gdsp"] Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.604686 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.607433 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.607491 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-dbh5q" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.620909 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-2gdsp"] Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.677025 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-15e4-account-create-ds56p"] Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.709358 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-db-sync-config-data\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.709408 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-combined-ca-bundle\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.709426 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9nkb\" (UniqueName: \"kubernetes.io/projected/24dc5231-a128-46dc-b0a2-5d13524ec5db-kube-api-access-s9nkb\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.709524 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-config-data\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.811517 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-config-data\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.811916 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-db-sync-config-data\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.811936 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-combined-ca-bundle\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.811957 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9nkb\" (UniqueName: \"kubernetes.io/projected/24dc5231-a128-46dc-b0a2-5d13524ec5db-kube-api-access-s9nkb\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.823987 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-db-sync-config-data\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.824182 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-config-data\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.824205 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-combined-ca-bundle\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.829740 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9nkb\" (UniqueName: \"kubernetes.io/projected/24dc5231-a128-46dc-b0a2-5d13524ec5db-kube-api-access-s9nkb\") pod \"watcher-db-sync-2gdsp\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:08 crc kubenswrapper[4810]: I0930 08:21:08.942724 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.009135 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" event={"ID":"cf25d291-3d0d-48c0-9dae-405ef9fd4922","Type":"ContainerStarted","Data":"11c46bda01fa7f7f62f2e761f2232d7700b63962f2aa93f556d70bd51276e6dd"} Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.009498 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.023692 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8-config-dvzf6" event={"ID":"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b","Type":"ContainerStarted","Data":"84151a80e112b5f2857d8d50b1a5e6ff3429de77a28966ba7d97d1da6b61f798"} Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.024471 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8-config-dvzf6" event={"ID":"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b","Type":"ContainerStarted","Data":"fb194c69b0b5c3c9a4f9d40971c0bfff8a4374a61bd15816a72a1ace2f9650a7"} Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.025856 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-15e4-account-create-ds56p" event={"ID":"fcd581d7-6a83-4822-822b-421633417bdf","Type":"ContainerStarted","Data":"065d5e95f2e03734ff22ed881a330c63409087875ce7e2d6d077ed8c07c48840"} Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.025909 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-15e4-account-create-ds56p" event={"ID":"fcd581d7-6a83-4822-822b-421633417bdf","Type":"ContainerStarted","Data":"c33ad2e5b72a7a24198e82f2026f9a258a27ba9d90aa40b517567e7c281a6169"} Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.038940 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" podStartSLOduration=3.038918796 podStartE2EDuration="3.038918796s" podCreationTimestamp="2025-09-30 08:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:21:09.033438347 +0000 UTC m=+1092.485637624" watchObservedRunningTime="2025-09-30 08:21:09.038918796 +0000 UTC m=+1092.491118063" Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.039113 4810 generic.go:334] "Generic (PLEG): container finished" podID="d505f4a8-0fe5-4cc5-9e55-369f42edd90a" containerID="47ab22daa69a39d7b3c6e9d6184d773f0f9535293c3381650bb88a69e4e68d81" exitCode=0 Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.039168 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dfa1-account-create-tnx8l" event={"ID":"d505f4a8-0fe5-4cc5-9e55-369f42edd90a","Type":"ContainerDied","Data":"47ab22daa69a39d7b3c6e9d6184d773f0f9535293c3381650bb88a69e4e68d81"} Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.039192 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dfa1-account-create-tnx8l" event={"ID":"d505f4a8-0fe5-4cc5-9e55-369f42edd90a","Type":"ContainerStarted","Data":"c69c8501aaf38975f10e8cce7330a59c2a5049161f934259eec8414296e3530e"} Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.041149 4810 generic.go:334] "Generic (PLEG): container finished" podID="127b94a4-4969-4a8d-b708-1b1c6f61d6af" containerID="7efd85c00404d4b93ee15bfe8912533c23ce84d7536d2e7ee22fde649024d43e" exitCode=0 Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.042998 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-588d-account-create-hl9zx" event={"ID":"127b94a4-4969-4a8d-b708-1b1c6f61d6af","Type":"ContainerDied","Data":"7efd85c00404d4b93ee15bfe8912533c23ce84d7536d2e7ee22fde649024d43e"} Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.043028 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-588d-account-create-hl9zx" event={"ID":"127b94a4-4969-4a8d-b708-1b1c6f61d6af","Type":"ContainerStarted","Data":"dd6fa11baee390f014380ff2b0ce56273fecbb95945ecd326f85e3197e36a0d1"} Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.054640 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-9q2d8-config-dvzf6" podStartSLOduration=2.054621773 podStartE2EDuration="2.054621773s" podCreationTimestamp="2025-09-30 08:21:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:21:09.047896157 +0000 UTC m=+1092.500095424" watchObservedRunningTime="2025-09-30 08:21:09.054621773 +0000 UTC m=+1092.506821040" Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.174715 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-410c-account-create-4czqf"] Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.318595 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1803bddd-901e-48bd-a1e9-50a3132e62e3" path="/var/lib/kubelet/pods/1803bddd-901e-48bd-a1e9-50a3132e62e3/volumes" Sep 30 08:21:09 crc kubenswrapper[4810]: I0930 08:21:09.528893 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-2gdsp"] Sep 30 08:21:09 crc kubenswrapper[4810]: W0930 08:21:09.538133 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24dc5231_a128_46dc_b0a2_5d13524ec5db.slice/crio-096294829df0fb36f43b016968d2016095288a6434167d8c3dba8f4f4c9afe80 WatchSource:0}: Error finding container 096294829df0fb36f43b016968d2016095288a6434167d8c3dba8f4f4c9afe80: Status 404 returned error can't find the container with id 096294829df0fb36f43b016968d2016095288a6434167d8c3dba8f4f4c9afe80 Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.051554 4810 generic.go:334] "Generic (PLEG): container finished" podID="ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b" containerID="84151a80e112b5f2857d8d50b1a5e6ff3429de77a28966ba7d97d1da6b61f798" exitCode=0 Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.051904 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8-config-dvzf6" event={"ID":"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b","Type":"ContainerDied","Data":"84151a80e112b5f2857d8d50b1a5e6ff3429de77a28966ba7d97d1da6b61f798"} Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.054176 4810 generic.go:334] "Generic (PLEG): container finished" podID="fcd581d7-6a83-4822-822b-421633417bdf" containerID="065d5e95f2e03734ff22ed881a330c63409087875ce7e2d6d077ed8c07c48840" exitCode=0 Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.054224 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-15e4-account-create-ds56p" event={"ID":"fcd581d7-6a83-4822-822b-421633417bdf","Type":"ContainerDied","Data":"065d5e95f2e03734ff22ed881a330c63409087875ce7e2d6d077ed8c07c48840"} Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.059576 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-2gdsp" event={"ID":"24dc5231-a128-46dc-b0a2-5d13524ec5db","Type":"ContainerStarted","Data":"096294829df0fb36f43b016968d2016095288a6434167d8c3dba8f4f4c9afe80"} Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.063100 4810 generic.go:334] "Generic (PLEG): container finished" podID="ca25948f-0522-43b1-9b6b-a7ce98207be0" containerID="ed4ab215943a49af7f1431e9a25e5b252f7be9e0092b220237217d4ba4095e24" exitCode=0 Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.063339 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-410c-account-create-4czqf" event={"ID":"ca25948f-0522-43b1-9b6b-a7ce98207be0","Type":"ContainerDied","Data":"ed4ab215943a49af7f1431e9a25e5b252f7be9e0092b220237217d4ba4095e24"} Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.063363 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-410c-account-create-4czqf" event={"ID":"ca25948f-0522-43b1-9b6b-a7ce98207be0","Type":"ContainerStarted","Data":"cb8442393cc3be971ea2e4d182695e51b3086f56e2713747aaa777d7c51b04fb"} Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.462806 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-588d-account-create-hl9zx" Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.569827 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmxwp\" (UniqueName: \"kubernetes.io/projected/127b94a4-4969-4a8d-b708-1b1c6f61d6af-kube-api-access-rmxwp\") pod \"127b94a4-4969-4a8d-b708-1b1c6f61d6af\" (UID: \"127b94a4-4969-4a8d-b708-1b1c6f61d6af\") " Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.576484 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/127b94a4-4969-4a8d-b708-1b1c6f61d6af-kube-api-access-rmxwp" (OuterVolumeSpecName: "kube-api-access-rmxwp") pod "127b94a4-4969-4a8d-b708-1b1c6f61d6af" (UID: "127b94a4-4969-4a8d-b708-1b1c6f61d6af"). InnerVolumeSpecName "kube-api-access-rmxwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.618545 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dfa1-account-create-tnx8l" Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.626223 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-15e4-account-create-ds56p" Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.672670 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmxwp\" (UniqueName: \"kubernetes.io/projected/127b94a4-4969-4a8d-b708-1b1c6f61d6af-kube-api-access-rmxwp\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.773657 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl7ct\" (UniqueName: \"kubernetes.io/projected/fcd581d7-6a83-4822-822b-421633417bdf-kube-api-access-bl7ct\") pod \"fcd581d7-6a83-4822-822b-421633417bdf\" (UID: \"fcd581d7-6a83-4822-822b-421633417bdf\") " Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.774091 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7tkr\" (UniqueName: \"kubernetes.io/projected/d505f4a8-0fe5-4cc5-9e55-369f42edd90a-kube-api-access-b7tkr\") pod \"d505f4a8-0fe5-4cc5-9e55-369f42edd90a\" (UID: \"d505f4a8-0fe5-4cc5-9e55-369f42edd90a\") " Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.777654 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcd581d7-6a83-4822-822b-421633417bdf-kube-api-access-bl7ct" (OuterVolumeSpecName: "kube-api-access-bl7ct") pod "fcd581d7-6a83-4822-822b-421633417bdf" (UID: "fcd581d7-6a83-4822-822b-421633417bdf"). InnerVolumeSpecName "kube-api-access-bl7ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.778133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d505f4a8-0fe5-4cc5-9e55-369f42edd90a-kube-api-access-b7tkr" (OuterVolumeSpecName: "kube-api-access-b7tkr") pod "d505f4a8-0fe5-4cc5-9e55-369f42edd90a" (UID: "d505f4a8-0fe5-4cc5-9e55-369f42edd90a"). InnerVolumeSpecName "kube-api-access-b7tkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.876088 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl7ct\" (UniqueName: \"kubernetes.io/projected/fcd581d7-6a83-4822-822b-421633417bdf-kube-api-access-bl7ct\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:10 crc kubenswrapper[4810]: I0930 08:21:10.876120 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7tkr\" (UniqueName: \"kubernetes.io/projected/d505f4a8-0fe5-4cc5-9e55-369f42edd90a-kube-api-access-b7tkr\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.079134 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dfa1-account-create-tnx8l" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.079440 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dfa1-account-create-tnx8l" event={"ID":"d505f4a8-0fe5-4cc5-9e55-369f42edd90a","Type":"ContainerDied","Data":"c69c8501aaf38975f10e8cce7330a59c2a5049161f934259eec8414296e3530e"} Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.079485 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c69c8501aaf38975f10e8cce7330a59c2a5049161f934259eec8414296e3530e" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.082038 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-588d-account-create-hl9zx" event={"ID":"127b94a4-4969-4a8d-b708-1b1c6f61d6af","Type":"ContainerDied","Data":"dd6fa11baee390f014380ff2b0ce56273fecbb95945ecd326f85e3197e36a0d1"} Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.082073 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd6fa11baee390f014380ff2b0ce56273fecbb95945ecd326f85e3197e36a0d1" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.082048 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-588d-account-create-hl9zx" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.083874 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-15e4-account-create-ds56p" event={"ID":"fcd581d7-6a83-4822-822b-421633417bdf","Type":"ContainerDied","Data":"c33ad2e5b72a7a24198e82f2026f9a258a27ba9d90aa40b517567e7c281a6169"} Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.083963 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c33ad2e5b72a7a24198e82f2026f9a258a27ba9d90aa40b517567e7c281a6169" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.083885 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-15e4-account-create-ds56p" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.214359 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5fa8-account-create-xtqgd"] Sep 30 08:21:11 crc kubenswrapper[4810]: E0930 08:21:11.214821 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcd581d7-6a83-4822-822b-421633417bdf" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.214837 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcd581d7-6a83-4822-822b-421633417bdf" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: E0930 08:21:11.215836 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d505f4a8-0fe5-4cc5-9e55-369f42edd90a" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.215862 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d505f4a8-0fe5-4cc5-9e55-369f42edd90a" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: E0930 08:21:11.215877 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127b94a4-4969-4a8d-b708-1b1c6f61d6af" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.215885 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="127b94a4-4969-4a8d-b708-1b1c6f61d6af" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.216096 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d505f4a8-0fe5-4cc5-9e55-369f42edd90a" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.216127 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="127b94a4-4969-4a8d-b708-1b1c6f61d6af" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.216156 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcd581d7-6a83-4822-822b-421633417bdf" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.216846 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5fa8-account-create-xtqgd" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.221175 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.233921 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5fa8-account-create-xtqgd"] Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.386554 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9vpg\" (UniqueName: \"kubernetes.io/projected/737b1bf9-6bc6-49a7-8913-5c78af3f1e82-kube-api-access-g9vpg\") pod \"keystone-5fa8-account-create-xtqgd\" (UID: \"737b1bf9-6bc6-49a7-8913-5c78af3f1e82\") " pod="openstack/keystone-5fa8-account-create-xtqgd" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.487989 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9vpg\" (UniqueName: \"kubernetes.io/projected/737b1bf9-6bc6-49a7-8913-5c78af3f1e82-kube-api-access-g9vpg\") pod \"keystone-5fa8-account-create-xtqgd\" (UID: \"737b1bf9-6bc6-49a7-8913-5c78af3f1e82\") " pod="openstack/keystone-5fa8-account-create-xtqgd" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.512528 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9vpg\" (UniqueName: \"kubernetes.io/projected/737b1bf9-6bc6-49a7-8913-5c78af3f1e82-kube-api-access-g9vpg\") pod \"keystone-5fa8-account-create-xtqgd\" (UID: \"737b1bf9-6bc6-49a7-8913-5c78af3f1e82\") " pod="openstack/keystone-5fa8-account-create-xtqgd" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.531615 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5fa8-account-create-xtqgd" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.600420 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-410c-account-create-4czqf" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.606738 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.691261 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gjzw\" (UniqueName: \"kubernetes.io/projected/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-kube-api-access-8gjzw\") pod \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.691346 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run\") pod \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.691377 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knpsj\" (UniqueName: \"kubernetes.io/projected/ca25948f-0522-43b1-9b6b-a7ce98207be0-kube-api-access-knpsj\") pod \"ca25948f-0522-43b1-9b6b-a7ce98207be0\" (UID: \"ca25948f-0522-43b1-9b6b-a7ce98207be0\") " Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.691425 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-log-ovn\") pod \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.691446 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-scripts\") pod \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.691475 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run-ovn\") pod \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.691495 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-additional-scripts\") pod \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\" (UID: \"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b\") " Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.691976 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b" (UID: "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.692046 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run" (OuterVolumeSpecName: "var-run") pod "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b" (UID: "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.693050 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b" (UID: "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.693809 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-scripts" (OuterVolumeSpecName: "scripts") pod "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b" (UID: "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.693842 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b" (UID: "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.697892 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca25948f-0522-43b1-9b6b-a7ce98207be0-kube-api-access-knpsj" (OuterVolumeSpecName: "kube-api-access-knpsj") pod "ca25948f-0522-43b1-9b6b-a7ce98207be0" (UID: "ca25948f-0522-43b1-9b6b-a7ce98207be0"). InnerVolumeSpecName "kube-api-access-knpsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.699373 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-kube-api-access-8gjzw" (OuterVolumeSpecName: "kube-api-access-8gjzw") pod "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b" (UID: "ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b"). InnerVolumeSpecName "kube-api-access-8gjzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.702424 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c962-account-create-swr7n"] Sep 30 08:21:11 crc kubenswrapper[4810]: E0930 08:21:11.702824 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca25948f-0522-43b1-9b6b-a7ce98207be0" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.702842 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca25948f-0522-43b1-9b6b-a7ce98207be0" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: E0930 08:21:11.702864 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b" containerName="ovn-config" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.702871 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b" containerName="ovn-config" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.703034 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b" containerName="ovn-config" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.703067 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca25948f-0522-43b1-9b6b-a7ce98207be0" containerName="mariadb-account-create" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.703664 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c962-account-create-swr7n" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.709211 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c962-account-create-swr7n"] Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.710891 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.793737 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9hbp\" (UniqueName: \"kubernetes.io/projected/8962a403-5d8b-4295-8c59-438ccf65fdbc-kube-api-access-f9hbp\") pod \"placement-c962-account-create-swr7n\" (UID: \"8962a403-5d8b-4295-8c59-438ccf65fdbc\") " pod="openstack/placement-c962-account-create-swr7n" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.794233 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.794261 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knpsj\" (UniqueName: \"kubernetes.io/projected/ca25948f-0522-43b1-9b6b-a7ce98207be0-kube-api-access-knpsj\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.794288 4810 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.794297 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.794305 4810 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.794313 4810 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-additional-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.794327 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gjzw\" (UniqueName: \"kubernetes.io/projected/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b-kube-api-access-8gjzw\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.895588 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9hbp\" (UniqueName: \"kubernetes.io/projected/8962a403-5d8b-4295-8c59-438ccf65fdbc-kube-api-access-f9hbp\") pod \"placement-c962-account-create-swr7n\" (UID: \"8962a403-5d8b-4295-8c59-438ccf65fdbc\") " pod="openstack/placement-c962-account-create-swr7n" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.913162 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9hbp\" (UniqueName: \"kubernetes.io/projected/8962a403-5d8b-4295-8c59-438ccf65fdbc-kube-api-access-f9hbp\") pod \"placement-c962-account-create-swr7n\" (UID: \"8962a403-5d8b-4295-8c59-438ccf65fdbc\") " pod="openstack/placement-c962-account-create-swr7n" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.972294 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4c6qh"] Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.974808 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.977766 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.977955 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-wjg2p" Sep 30 08:21:11 crc kubenswrapper[4810]: I0930 08:21:11.986767 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4c6qh"] Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.015048 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5fa8-account-create-xtqgd"] Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.024509 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c962-account-create-swr7n" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.098682 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqpxq\" (UniqueName: \"kubernetes.io/projected/ff6ce927-f393-4895-a7b6-af266138c8c1-kube-api-access-vqpxq\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.098823 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-combined-ca-bundle\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.098857 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-config-data\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.098886 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-db-sync-config-data\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.100133 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9q2d8-config-dvzf6" event={"ID":"ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b","Type":"ContainerDied","Data":"fb194c69b0b5c3c9a4f9d40971c0bfff8a4374a61bd15816a72a1ace2f9650a7"} Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.100189 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb194c69b0b5c3c9a4f9d40971c0bfff8a4374a61bd15816a72a1ace2f9650a7" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.100148 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9q2d8-config-dvzf6" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.101936 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5fa8-account-create-xtqgd" event={"ID":"737b1bf9-6bc6-49a7-8913-5c78af3f1e82","Type":"ContainerStarted","Data":"50cf7e16609482fe6fc5aadc0769303b1e19e943f3c2910fd14b2d2e7ee7e26e"} Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.105199 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-410c-account-create-4czqf" event={"ID":"ca25948f-0522-43b1-9b6b-a7ce98207be0","Type":"ContainerDied","Data":"cb8442393cc3be971ea2e4d182695e51b3086f56e2713747aaa777d7c51b04fb"} Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.105249 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb8442393cc3be971ea2e4d182695e51b3086f56e2713747aaa777d7c51b04fb" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.105321 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-410c-account-create-4czqf" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.142597 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-9q2d8-config-dvzf6"] Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.154494 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-9q2d8-config-dvzf6"] Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.200486 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-combined-ca-bundle\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.200529 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-config-data\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.200555 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-db-sync-config-data\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.200646 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqpxq\" (UniqueName: \"kubernetes.io/projected/ff6ce927-f393-4895-a7b6-af266138c8c1-kube-api-access-vqpxq\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.207851 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-db-sync-config-data\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.208296 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-combined-ca-bundle\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.212015 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-config-data\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.224701 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqpxq\" (UniqueName: \"kubernetes.io/projected/ff6ce927-f393-4895-a7b6-af266138c8c1-kube-api-access-vqpxq\") pod \"glance-db-sync-4c6qh\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.295014 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4c6qh" Sep 30 08:21:12 crc kubenswrapper[4810]: I0930 08:21:12.508980 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c962-account-create-swr7n"] Sep 30 08:21:13 crc kubenswrapper[4810]: I0930 08:21:13.114724 4810 generic.go:334] "Generic (PLEG): container finished" podID="eb9bba80-4046-4788-ac20-c682625132d8" containerID="43a44738cda6d3139fcbce99404f6f250153b2205c24b1b2cda00b30fe0cb1ff" exitCode=0 Sep 30 08:21:13 crc kubenswrapper[4810]: I0930 08:21:13.114815 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb9bba80-4046-4788-ac20-c682625132d8","Type":"ContainerDied","Data":"43a44738cda6d3139fcbce99404f6f250153b2205c24b1b2cda00b30fe0cb1ff"} Sep 30 08:21:13 crc kubenswrapper[4810]: I0930 08:21:13.117416 4810 generic.go:334] "Generic (PLEG): container finished" podID="737b1bf9-6bc6-49a7-8913-5c78af3f1e82" containerID="bd85fffad441a4190e82453ea8a6577c502fee4b917a4cb4f2ebac02d71c9e73" exitCode=0 Sep 30 08:21:13 crc kubenswrapper[4810]: I0930 08:21:13.117459 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5fa8-account-create-xtqgd" event={"ID":"737b1bf9-6bc6-49a7-8913-5c78af3f1e82","Type":"ContainerDied","Data":"bd85fffad441a4190e82453ea8a6577c502fee4b917a4cb4f2ebac02d71c9e73"} Sep 30 08:21:13 crc kubenswrapper[4810]: I0930 08:21:13.323692 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b" path="/var/lib/kubelet/pods/ea9ba09a-2f1a-4a73-b0f1-c9ec3777806b/volumes" Sep 30 08:21:15 crc kubenswrapper[4810]: E0930 08:21:15.035005 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dacf67c_34b5_4c58_9f8d_00adfdcf74b7.slice\": RecentStats: unable to find data in memory cache]" Sep 30 08:21:16 crc kubenswrapper[4810]: I0930 08:21:16.424910 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5fa8-account-create-xtqgd" Sep 30 08:21:16 crc kubenswrapper[4810]: I0930 08:21:16.490216 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9vpg\" (UniqueName: \"kubernetes.io/projected/737b1bf9-6bc6-49a7-8913-5c78af3f1e82-kube-api-access-g9vpg\") pod \"737b1bf9-6bc6-49a7-8913-5c78af3f1e82\" (UID: \"737b1bf9-6bc6-49a7-8913-5c78af3f1e82\") " Sep 30 08:21:16 crc kubenswrapper[4810]: I0930 08:21:16.497479 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/737b1bf9-6bc6-49a7-8913-5c78af3f1e82-kube-api-access-g9vpg" (OuterVolumeSpecName: "kube-api-access-g9vpg") pod "737b1bf9-6bc6-49a7-8913-5c78af3f1e82" (UID: "737b1bf9-6bc6-49a7-8913-5c78af3f1e82"). InnerVolumeSpecName "kube-api-access-g9vpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:16 crc kubenswrapper[4810]: I0930 08:21:16.564452 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:21:16 crc kubenswrapper[4810]: I0930 08:21:16.592592 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9vpg\" (UniqueName: \"kubernetes.io/projected/737b1bf9-6bc6-49a7-8913-5c78af3f1e82-kube-api-access-g9vpg\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:16 crc kubenswrapper[4810]: I0930 08:21:16.628843 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c8b4c69d9-t2bvs"] Sep 30 08:21:16 crc kubenswrapper[4810]: I0930 08:21:16.629110 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" podUID="73deb073-9ff6-4ef6-8e08-98ecc936dca1" containerName="dnsmasq-dns" containerID="cri-o://eba10c72387b210fe793bfada9c89f53eecbcaeb43a88bbfd9ccd9eaa76376d3" gracePeriod=10 Sep 30 08:21:17 crc kubenswrapper[4810]: I0930 08:21:17.163208 4810 generic.go:334] "Generic (PLEG): container finished" podID="73deb073-9ff6-4ef6-8e08-98ecc936dca1" containerID="eba10c72387b210fe793bfada9c89f53eecbcaeb43a88bbfd9ccd9eaa76376d3" exitCode=0 Sep 30 08:21:17 crc kubenswrapper[4810]: I0930 08:21:17.163345 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" event={"ID":"73deb073-9ff6-4ef6-8e08-98ecc936dca1","Type":"ContainerDied","Data":"eba10c72387b210fe793bfada9c89f53eecbcaeb43a88bbfd9ccd9eaa76376d3"} Sep 30 08:21:17 crc kubenswrapper[4810]: I0930 08:21:17.165678 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5fa8-account-create-xtqgd" event={"ID":"737b1bf9-6bc6-49a7-8913-5c78af3f1e82","Type":"ContainerDied","Data":"50cf7e16609482fe6fc5aadc0769303b1e19e943f3c2910fd14b2d2e7ee7e26e"} Sep 30 08:21:17 crc kubenswrapper[4810]: I0930 08:21:17.165719 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50cf7e16609482fe6fc5aadc0769303b1e19e943f3c2910fd14b2d2e7ee7e26e" Sep 30 08:21:17 crc kubenswrapper[4810]: I0930 08:21:17.165776 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5fa8-account-create-xtqgd" Sep 30 08:21:17 crc kubenswrapper[4810]: W0930 08:21:17.524611 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8962a403_5d8b_4295_8c59_438ccf65fdbc.slice/crio-db057b08ac730b8a236d58b26e0885492427c855aaebca26fd8dc8e2f7321c27 WatchSource:0}: Error finding container db057b08ac730b8a236d58b26e0885492427c855aaebca26fd8dc8e2f7321c27: Status 404 returned error can't find the container with id db057b08ac730b8a236d58b26e0885492427c855aaebca26fd8dc8e2f7321c27 Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.182534 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb9bba80-4046-4788-ac20-c682625132d8","Type":"ContainerStarted","Data":"66e23df6841865f3e52782074053a40ffa2afa2c03e76c69473fa0655c6d52f3"} Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.187403 4810 generic.go:334] "Generic (PLEG): container finished" podID="8962a403-5d8b-4295-8c59-438ccf65fdbc" containerID="b0373a01d538c3646517d116acc2f25bed8ebfbd8ad9eb70be0f142464ef077b" exitCode=0 Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.187486 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c962-account-create-swr7n" event={"ID":"8962a403-5d8b-4295-8c59-438ccf65fdbc","Type":"ContainerDied","Data":"b0373a01d538c3646517d116acc2f25bed8ebfbd8ad9eb70be0f142464ef077b"} Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.187519 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c962-account-create-swr7n" event={"ID":"8962a403-5d8b-4295-8c59-438ccf65fdbc","Type":"ContainerStarted","Data":"db057b08ac730b8a236d58b26e0885492427c855aaebca26fd8dc8e2f7321c27"} Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.189208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-2gdsp" event={"ID":"24dc5231-a128-46dc-b0a2-5d13524ec5db","Type":"ContainerStarted","Data":"e81c0d5ff5a09a1428e069d6c9847bd00a288bad6b4ac403b71e606e22914f9e"} Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.199832 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.202537 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4c6qh"] Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.223061 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-2gdsp" podStartSLOduration=2.052328253 podStartE2EDuration="10.223040498s" podCreationTimestamp="2025-09-30 08:21:08 +0000 UTC" firstStartedPulling="2025-09-30 08:21:09.539763739 +0000 UTC m=+1092.991963016" lastFinishedPulling="2025-09-30 08:21:17.710475994 +0000 UTC m=+1101.162675261" observedRunningTime="2025-09-30 08:21:18.222739799 +0000 UTC m=+1101.674939066" watchObservedRunningTime="2025-09-30 08:21:18.223040498 +0000 UTC m=+1101.675239765" Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.332924 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-config\") pod \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.332978 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-nb\") pod \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.333008 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-sb\") pod \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.333057 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dgtb\" (UniqueName: \"kubernetes.io/projected/73deb073-9ff6-4ef6-8e08-98ecc936dca1-kube-api-access-8dgtb\") pod \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.333185 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-dns-svc\") pod \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\" (UID: \"73deb073-9ff6-4ef6-8e08-98ecc936dca1\") " Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.337002 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73deb073-9ff6-4ef6-8e08-98ecc936dca1-kube-api-access-8dgtb" (OuterVolumeSpecName: "kube-api-access-8dgtb") pod "73deb073-9ff6-4ef6-8e08-98ecc936dca1" (UID: "73deb073-9ff6-4ef6-8e08-98ecc936dca1"). InnerVolumeSpecName "kube-api-access-8dgtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.380292 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "73deb073-9ff6-4ef6-8e08-98ecc936dca1" (UID: "73deb073-9ff6-4ef6-8e08-98ecc936dca1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.380705 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "73deb073-9ff6-4ef6-8e08-98ecc936dca1" (UID: "73deb073-9ff6-4ef6-8e08-98ecc936dca1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.387588 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "73deb073-9ff6-4ef6-8e08-98ecc936dca1" (UID: "73deb073-9ff6-4ef6-8e08-98ecc936dca1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.392650 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-config" (OuterVolumeSpecName: "config") pod "73deb073-9ff6-4ef6-8e08-98ecc936dca1" (UID: "73deb073-9ff6-4ef6-8e08-98ecc936dca1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.440204 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.440247 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dgtb\" (UniqueName: \"kubernetes.io/projected/73deb073-9ff6-4ef6-8e08-98ecc936dca1-kube-api-access-8dgtb\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.440263 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.440293 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:18 crc kubenswrapper[4810]: I0930 08:21:18.440306 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73deb073-9ff6-4ef6-8e08-98ecc936dca1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.197913 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4c6qh" event={"ID":"ff6ce927-f393-4895-a7b6-af266138c8c1","Type":"ContainerStarted","Data":"0760ebf1f31bcdf28e93b3d14f8865b225951455df4408d3ee32e89338d0e68e"} Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.200298 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" event={"ID":"73deb073-9ff6-4ef6-8e08-98ecc936dca1","Type":"ContainerDied","Data":"0ce0d454fe6cc9ba1fd245afcc8a68433f89a31289824e3a69f79e30039ac7e2"} Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.200383 4810 scope.go:117] "RemoveContainer" containerID="eba10c72387b210fe793bfada9c89f53eecbcaeb43a88bbfd9ccd9eaa76376d3" Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.200586 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8b4c69d9-t2bvs" Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.228246 4810 scope.go:117] "RemoveContainer" containerID="fb2fc73b40d7f34790ae4b74b80e5d66fd77a31ddde5e3c943915c0bd78f339e" Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.232694 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c8b4c69d9-t2bvs"] Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.240510 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c8b4c69d9-t2bvs"] Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.320830 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73deb073-9ff6-4ef6-8e08-98ecc936dca1" path="/var/lib/kubelet/pods/73deb073-9ff6-4ef6-8e08-98ecc936dca1/volumes" Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.594789 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c962-account-create-swr7n" Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.660482 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9hbp\" (UniqueName: \"kubernetes.io/projected/8962a403-5d8b-4295-8c59-438ccf65fdbc-kube-api-access-f9hbp\") pod \"8962a403-5d8b-4295-8c59-438ccf65fdbc\" (UID: \"8962a403-5d8b-4295-8c59-438ccf65fdbc\") " Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.687238 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8962a403-5d8b-4295-8c59-438ccf65fdbc-kube-api-access-f9hbp" (OuterVolumeSpecName: "kube-api-access-f9hbp") pod "8962a403-5d8b-4295-8c59-438ccf65fdbc" (UID: "8962a403-5d8b-4295-8c59-438ccf65fdbc"). InnerVolumeSpecName "kube-api-access-f9hbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:19 crc kubenswrapper[4810]: I0930 08:21:19.762195 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9hbp\" (UniqueName: \"kubernetes.io/projected/8962a403-5d8b-4295-8c59-438ccf65fdbc-kube-api-access-f9hbp\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:20 crc kubenswrapper[4810]: I0930 08:21:20.211079 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c962-account-create-swr7n" event={"ID":"8962a403-5d8b-4295-8c59-438ccf65fdbc","Type":"ContainerDied","Data":"db057b08ac730b8a236d58b26e0885492427c855aaebca26fd8dc8e2f7321c27"} Sep 30 08:21:20 crc kubenswrapper[4810]: I0930 08:21:20.211340 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db057b08ac730b8a236d58b26e0885492427c855aaebca26fd8dc8e2f7321c27" Sep 30 08:21:20 crc kubenswrapper[4810]: I0930 08:21:20.211108 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c962-account-create-swr7n" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.222603 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb9bba80-4046-4788-ac20-c682625132d8","Type":"ContainerStarted","Data":"612a628dc5694fefa9476ba97ed57198bc80e617d9dbe0dcfb09e14d0b996d8d"} Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.222644 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb9bba80-4046-4788-ac20-c682625132d8","Type":"ContainerStarted","Data":"a274e25cded7421e038e2251ec27af7dbe8172d4c7e4d2c06168aa669836f3d0"} Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.264357 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=18.264333287 podStartE2EDuration="18.264333287s" podCreationTimestamp="2025-09-30 08:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:21:21.250179556 +0000 UTC m=+1104.702378823" watchObservedRunningTime="2025-09-30 08:21:21.264333287 +0000 UTC m=+1104.716532554" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.798254 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xd7mq"] Sep 30 08:21:21 crc kubenswrapper[4810]: E0930 08:21:21.798582 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73deb073-9ff6-4ef6-8e08-98ecc936dca1" containerName="dnsmasq-dns" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.798599 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="73deb073-9ff6-4ef6-8e08-98ecc936dca1" containerName="dnsmasq-dns" Sep 30 08:21:21 crc kubenswrapper[4810]: E0930 08:21:21.798615 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8962a403-5d8b-4295-8c59-438ccf65fdbc" containerName="mariadb-account-create" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.798623 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8962a403-5d8b-4295-8c59-438ccf65fdbc" containerName="mariadb-account-create" Sep 30 08:21:21 crc kubenswrapper[4810]: E0930 08:21:21.798633 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="737b1bf9-6bc6-49a7-8913-5c78af3f1e82" containerName="mariadb-account-create" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.798649 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="737b1bf9-6bc6-49a7-8913-5c78af3f1e82" containerName="mariadb-account-create" Sep 30 08:21:21 crc kubenswrapper[4810]: E0930 08:21:21.798672 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73deb073-9ff6-4ef6-8e08-98ecc936dca1" containerName="init" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.798679 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="73deb073-9ff6-4ef6-8e08-98ecc936dca1" containerName="init" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.798826 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="737b1bf9-6bc6-49a7-8913-5c78af3f1e82" containerName="mariadb-account-create" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.798846 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="73deb073-9ff6-4ef6-8e08-98ecc936dca1" containerName="dnsmasq-dns" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.798864 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8962a403-5d8b-4295-8c59-438ccf65fdbc" containerName="mariadb-account-create" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.799372 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.801369 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-hw6qv" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.802204 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.804305 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.805336 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.816136 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xd7mq"] Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.899584 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-config-data\") pod \"keystone-db-sync-xd7mq\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.900225 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcr5f\" (UniqueName: \"kubernetes.io/projected/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-kube-api-access-dcr5f\") pod \"keystone-db-sync-xd7mq\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:21 crc kubenswrapper[4810]: I0930 08:21:21.900391 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-combined-ca-bundle\") pod \"keystone-db-sync-xd7mq\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:22 crc kubenswrapper[4810]: I0930 08:21:22.004499 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-config-data\") pod \"keystone-db-sync-xd7mq\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:22 crc kubenswrapper[4810]: I0930 08:21:22.005176 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcr5f\" (UniqueName: \"kubernetes.io/projected/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-kube-api-access-dcr5f\") pod \"keystone-db-sync-xd7mq\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:22 crc kubenswrapper[4810]: I0930 08:21:22.005284 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-combined-ca-bundle\") pod \"keystone-db-sync-xd7mq\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:22 crc kubenswrapper[4810]: I0930 08:21:22.014077 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-config-data\") pod \"keystone-db-sync-xd7mq\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:22 crc kubenswrapper[4810]: I0930 08:21:22.014638 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-combined-ca-bundle\") pod \"keystone-db-sync-xd7mq\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:22 crc kubenswrapper[4810]: I0930 08:21:22.024046 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcr5f\" (UniqueName: \"kubernetes.io/projected/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-kube-api-access-dcr5f\") pod \"keystone-db-sync-xd7mq\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:22 crc kubenswrapper[4810]: I0930 08:21:22.116498 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:22 crc kubenswrapper[4810]: I0930 08:21:22.236047 4810 generic.go:334] "Generic (PLEG): container finished" podID="24dc5231-a128-46dc-b0a2-5d13524ec5db" containerID="e81c0d5ff5a09a1428e069d6c9847bd00a288bad6b4ac403b71e606e22914f9e" exitCode=0 Sep 30 08:21:22 crc kubenswrapper[4810]: I0930 08:21:22.237628 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-2gdsp" event={"ID":"24dc5231-a128-46dc-b0a2-5d13524ec5db","Type":"ContainerDied","Data":"e81c0d5ff5a09a1428e069d6c9847bd00a288bad6b4ac403b71e606e22914f9e"} Sep 30 08:21:22 crc kubenswrapper[4810]: I0930 08:21:22.635365 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xd7mq"] Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.248306 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xd7mq" event={"ID":"df504df8-0ce9-4250-9ffc-db3f3e1ee26e","Type":"ContainerStarted","Data":"8680c4cb7a7ec9ad9c1d00e607c30aad6c4f6310d3f4d933a3797f1f32e56a89"} Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.622453 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.734080 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-db-sync-config-data\") pod \"24dc5231-a128-46dc-b0a2-5d13524ec5db\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.734248 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-combined-ca-bundle\") pod \"24dc5231-a128-46dc-b0a2-5d13524ec5db\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.734347 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-config-data\") pod \"24dc5231-a128-46dc-b0a2-5d13524ec5db\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.735397 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9nkb\" (UniqueName: \"kubernetes.io/projected/24dc5231-a128-46dc-b0a2-5d13524ec5db-kube-api-access-s9nkb\") pod \"24dc5231-a128-46dc-b0a2-5d13524ec5db\" (UID: \"24dc5231-a128-46dc-b0a2-5d13524ec5db\") " Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.740156 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24dc5231-a128-46dc-b0a2-5d13524ec5db-kube-api-access-s9nkb" (OuterVolumeSpecName: "kube-api-access-s9nkb") pod "24dc5231-a128-46dc-b0a2-5d13524ec5db" (UID: "24dc5231-a128-46dc-b0a2-5d13524ec5db"). InnerVolumeSpecName "kube-api-access-s9nkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.741182 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "24dc5231-a128-46dc-b0a2-5d13524ec5db" (UID: "24dc5231-a128-46dc-b0a2-5d13524ec5db"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.753374 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.776496 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24dc5231-a128-46dc-b0a2-5d13524ec5db" (UID: "24dc5231-a128-46dc-b0a2-5d13524ec5db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.787015 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-config-data" (OuterVolumeSpecName: "config-data") pod "24dc5231-a128-46dc-b0a2-5d13524ec5db" (UID: "24dc5231-a128-46dc-b0a2-5d13524ec5db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.838252 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.838307 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.838321 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9nkb\" (UniqueName: \"kubernetes.io/projected/24dc5231-a128-46dc-b0a2-5d13524ec5db-kube-api-access-s9nkb\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:23 crc kubenswrapper[4810]: I0930 08:21:23.838335 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/24dc5231-a128-46dc-b0a2-5d13524ec5db-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:24 crc kubenswrapper[4810]: I0930 08:21:24.258605 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-2gdsp" event={"ID":"24dc5231-a128-46dc-b0a2-5d13524ec5db","Type":"ContainerDied","Data":"096294829df0fb36f43b016968d2016095288a6434167d8c3dba8f4f4c9afe80"} Sep 30 08:21:24 crc kubenswrapper[4810]: I0930 08:21:24.258878 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="096294829df0fb36f43b016968d2016095288a6434167d8c3dba8f4f4c9afe80" Sep 30 08:21:24 crc kubenswrapper[4810]: I0930 08:21:24.258691 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-2gdsp" Sep 30 08:21:25 crc kubenswrapper[4810]: E0930 08:21:25.258308 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dacf67c_34b5_4c58_9f8d_00adfdcf74b7.slice\": RecentStats: unable to find data in memory cache]" Sep 30 08:21:33 crc kubenswrapper[4810]: I0930 08:21:33.755807 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:33 crc kubenswrapper[4810]: I0930 08:21:33.783589 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:34 crc kubenswrapper[4810]: I0930 08:21:34.359300 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Sep 30 08:21:35 crc kubenswrapper[4810]: E0930 08:21:35.507682 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dacf67c_34b5_4c58_9f8d_00adfdcf74b7.slice\": RecentStats: unable to find data in memory cache]" Sep 30 08:21:36 crc kubenswrapper[4810]: E0930 08:21:36.188564 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Sep 30 08:21:36 crc kubenswrapper[4810]: E0930 08:21:36.188637 4810 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Sep 30 08:21:36 crc kubenswrapper[4810]: E0930 08:21:36.188884 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:38.102.83.41:5001/podified-master-centos10/openstack-glance-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vqpxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-4c6qh_openstack(ff6ce927-f393-4895-a7b6-af266138c8c1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 08:21:36 crc kubenswrapper[4810]: E0930 08:21:36.190132 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-4c6qh" podUID="ff6ce927-f393-4895-a7b6-af266138c8c1" Sep 30 08:21:36 crc kubenswrapper[4810]: E0930 08:21:36.376966 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.41:5001/podified-master-centos10/openstack-glance-api:watcher_latest\\\"\"" pod="openstack/glance-db-sync-4c6qh" podUID="ff6ce927-f393-4895-a7b6-af266138c8c1" Sep 30 08:21:37 crc kubenswrapper[4810]: I0930 08:21:37.390509 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xd7mq" event={"ID":"df504df8-0ce9-4250-9ffc-db3f3e1ee26e","Type":"ContainerStarted","Data":"03579d2144c26fe931742e864142471db253a179541dd1e4eea42c5429bab5b8"} Sep 30 08:21:37 crc kubenswrapper[4810]: I0930 08:21:37.429475 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xd7mq" podStartSLOduration=2.913723311 podStartE2EDuration="16.429441925s" podCreationTimestamp="2025-09-30 08:21:21 +0000 UTC" firstStartedPulling="2025-09-30 08:21:22.660541657 +0000 UTC m=+1106.112740924" lastFinishedPulling="2025-09-30 08:21:36.176260231 +0000 UTC m=+1119.628459538" observedRunningTime="2025-09-30 08:21:37.413006738 +0000 UTC m=+1120.865206045" watchObservedRunningTime="2025-09-30 08:21:37.429441925 +0000 UTC m=+1120.881641212" Sep 30 08:21:40 crc kubenswrapper[4810]: I0930 08:21:40.425788 4810 generic.go:334] "Generic (PLEG): container finished" podID="df504df8-0ce9-4250-9ffc-db3f3e1ee26e" containerID="03579d2144c26fe931742e864142471db253a179541dd1e4eea42c5429bab5b8" exitCode=0 Sep 30 08:21:40 crc kubenswrapper[4810]: I0930 08:21:40.426144 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xd7mq" event={"ID":"df504df8-0ce9-4250-9ffc-db3f3e1ee26e","Type":"ContainerDied","Data":"03579d2144c26fe931742e864142471db253a179541dd1e4eea42c5429bab5b8"} Sep 30 08:21:41 crc kubenswrapper[4810]: I0930 08:21:41.958731 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.038537 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcr5f\" (UniqueName: \"kubernetes.io/projected/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-kube-api-access-dcr5f\") pod \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.038945 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-combined-ca-bundle\") pod \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.039037 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-config-data\") pod \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\" (UID: \"df504df8-0ce9-4250-9ffc-db3f3e1ee26e\") " Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.065828 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-kube-api-access-dcr5f" (OuterVolumeSpecName: "kube-api-access-dcr5f") pod "df504df8-0ce9-4250-9ffc-db3f3e1ee26e" (UID: "df504df8-0ce9-4250-9ffc-db3f3e1ee26e"). InnerVolumeSpecName "kube-api-access-dcr5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.087534 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df504df8-0ce9-4250-9ffc-db3f3e1ee26e" (UID: "df504df8-0ce9-4250-9ffc-db3f3e1ee26e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.093913 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-config-data" (OuterVolumeSpecName: "config-data") pod "df504df8-0ce9-4250-9ffc-db3f3e1ee26e" (UID: "df504df8-0ce9-4250-9ffc-db3f3e1ee26e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.140899 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcr5f\" (UniqueName: \"kubernetes.io/projected/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-kube-api-access-dcr5f\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.140942 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.140953 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df504df8-0ce9-4250-9ffc-db3f3e1ee26e-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.463482 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xd7mq" event={"ID":"df504df8-0ce9-4250-9ffc-db3f3e1ee26e","Type":"ContainerDied","Data":"8680c4cb7a7ec9ad9c1d00e607c30aad6c4f6310d3f4d933a3797f1f32e56a89"} Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.463547 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8680c4cb7a7ec9ad9c1d00e607c30aad6c4f6310d3f4d933a3797f1f32e56a89" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.463633 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xd7mq" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.928761 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-8xbnb"] Sep 30 08:21:42 crc kubenswrapper[4810]: E0930 08:21:42.940456 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24dc5231-a128-46dc-b0a2-5d13524ec5db" containerName="watcher-db-sync" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.940494 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="24dc5231-a128-46dc-b0a2-5d13524ec5db" containerName="watcher-db-sync" Sep 30 08:21:42 crc kubenswrapper[4810]: E0930 08:21:42.940506 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df504df8-0ce9-4250-9ffc-db3f3e1ee26e" containerName="keystone-db-sync" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.940512 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="df504df8-0ce9-4250-9ffc-db3f3e1ee26e" containerName="keystone-db-sync" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.940761 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="24dc5231-a128-46dc-b0a2-5d13524ec5db" containerName="watcher-db-sync" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.940785 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="df504df8-0ce9-4250-9ffc-db3f3e1ee26e" containerName="keystone-db-sync" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.941390 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.948116 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b74bb6999-gs5cq"] Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.949263 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.978350 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8xbnb"] Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.981609 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.991672 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.991876 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.991986 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-hw6qv" Sep 30 08:21:42 crc kubenswrapper[4810]: I0930 08:21:42.998773 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b74bb6999-gs5cq"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.072323 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.073494 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082135 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-scripts\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082176 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-svc\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082200 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sml94\" (UniqueName: \"kubernetes.io/projected/8ed92924-5cd6-4d38-ac53-7db3c0773437-kube-api-access-sml94\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082228 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-config\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082293 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-fernet-keys\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082324 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082342 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082364 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082379 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-combined-ca-bundle\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082400 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-config-data\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082430 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-credential-keys\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.082474 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbfnt\" (UniqueName: \"kubernetes.io/projected/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-kube-api-access-xbfnt\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.097334 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.110513 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.110574 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-dbh5q" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.183754 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/277a70ee-91dd-4198-9abd-7e6bb909bec0-logs\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.183810 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-scripts\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.183831 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-svc\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.183849 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sml94\" (UniqueName: \"kubernetes.io/projected/8ed92924-5cd6-4d38-ac53-7db3c0773437-kube-api-access-sml94\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.183875 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-config-data\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.183909 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-config\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.183967 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-fernet-keys\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.183994 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59pdh\" (UniqueName: \"kubernetes.io/projected/277a70ee-91dd-4198-9abd-7e6bb909bec0-kube-api-access-59pdh\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.184016 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.184034 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.184052 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.184069 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-combined-ca-bundle\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.184084 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-config-data\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.184115 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-credential-keys\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.184135 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.184172 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbfnt\" (UniqueName: \"kubernetes.io/projected/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-kube-api-access-xbfnt\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.185319 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.189517 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-scripts\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.190883 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.191176 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.191974 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-svc\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.193581 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-config-data\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.195576 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.195739 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-config\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.197592 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.199369 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-fernet-keys\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.204256 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.205648 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.208851 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.210327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-credential-keys\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.217520 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-combined-ca-bundle\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.218171 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.242534 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.246233 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sml94\" (UniqueName: \"kubernetes.io/projected/8ed92924-5cd6-4d38-ac53-7db3c0773437-kube-api-access-sml94\") pod \"keystone-bootstrap-8xbnb\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.256545 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-z7k2l"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.262300 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.265249 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.266653 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-lkx2n" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.275330 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.275716 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287378 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-logs\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287438 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc876eb0-0451-496a-90c1-4470a9d21ca2-logs\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287465 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287503 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59pdh\" (UniqueName: \"kubernetes.io/projected/277a70ee-91dd-4198-9abd-7e6bb909bec0-kube-api-access-59pdh\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287518 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-config-data\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287540 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-config-data\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287756 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287802 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287850 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287913 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/277a70ee-91dd-4198-9abd-7e6bb909bec0-logs\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287951 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlkbq\" (UniqueName: \"kubernetes.io/projected/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-kube-api-access-nlkbq\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287970 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dkwf\" (UniqueName: \"kubernetes.io/projected/bc876eb0-0451-496a-90c1-4470a9d21ca2-kube-api-access-5dkwf\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.287987 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-config-data\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.288015 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.288757 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/277a70ee-91dd-4198-9abd-7e6bb909bec0-logs\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.292132 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.292181 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.293633 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbfnt\" (UniqueName: \"kubernetes.io/projected/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-kube-api-access-xbfnt\") pod \"dnsmasq-dns-6b74bb6999-gs5cq\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.294177 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-config-data\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.333711 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5648749bd9-8rckr"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.337846 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.349776 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-9w7ds" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.350017 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.350137 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.350319 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.357993 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59pdh\" (UniqueName: \"kubernetes.io/projected/277a70ee-91dd-4198-9abd-7e6bb909bec0-kube-api-access-59pdh\") pod \"watcher-applier-0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391038 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-scripts\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391088 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-config-data\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391133 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-config-data\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391177 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-config-data\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391223 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391249 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e199d2a8-2a58-47c7-8018-13b29c37c2ad-etc-machine-id\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391288 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-db-sync-config-data\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391315 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-combined-ca-bundle\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391340 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ngd6\" (UniqueName: \"kubernetes.io/projected/4bf21836-6124-4093-89ce-5512a9239d73-kube-api-access-8ngd6\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391363 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391387 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4bf21836-6124-4093-89ce-5512a9239d73-horizon-secret-key\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391435 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlkbq\" (UniqueName: \"kubernetes.io/projected/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-kube-api-access-nlkbq\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391455 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dkwf\" (UniqueName: \"kubernetes.io/projected/bc876eb0-0451-496a-90c1-4470a9d21ca2-kube-api-access-5dkwf\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391481 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bf21836-6124-4093-89ce-5512a9239d73-logs\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391506 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391557 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qggnm\" (UniqueName: \"kubernetes.io/projected/e199d2a8-2a58-47c7-8018-13b29c37c2ad-kube-api-access-qggnm\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391578 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-logs\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391597 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc876eb0-0451-496a-90c1-4470a9d21ca2-logs\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391613 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391635 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-scripts\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.391651 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-config-data\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.395485 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.405666 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc876eb0-0451-496a-90c1-4470a9d21ca2-logs\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.406018 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-logs\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.406651 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-config-data\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.413913 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.416961 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.417029 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-z7k2l"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.429835 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.430515 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.431032 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-config-data\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.450304 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5648749bd9-8rckr"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.450948 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlkbq\" (UniqueName: \"kubernetes.io/projected/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-kube-api-access-nlkbq\") pod \"watcher-decision-engine-0\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.477607 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.480103 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dkwf\" (UniqueName: \"kubernetes.io/projected/bc876eb0-0451-496a-90c1-4470a9d21ca2-kube-api-access-5dkwf\") pod \"watcher-api-0\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.495537 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bf21836-6124-4093-89ce-5512a9239d73-logs\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.495608 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qggnm\" (UniqueName: \"kubernetes.io/projected/e199d2a8-2a58-47c7-8018-13b29c37c2ad-kube-api-access-qggnm\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.495644 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-scripts\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.495663 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-config-data\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.495685 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-scripts\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.495718 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-config-data\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.495749 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-db-sync-config-data\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.495765 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e199d2a8-2a58-47c7-8018-13b29c37c2ad-etc-machine-id\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.495789 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-combined-ca-bundle\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.495813 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ngd6\" (UniqueName: \"kubernetes.io/projected/4bf21836-6124-4093-89ce-5512a9239d73-kube-api-access-8ngd6\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.495839 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4bf21836-6124-4093-89ce-5512a9239d73-horizon-secret-key\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.500530 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-scripts\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.500794 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bf21836-6124-4093-89ce-5512a9239d73-logs\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.505868 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e199d2a8-2a58-47c7-8018-13b29c37c2ad-etc-machine-id\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.506815 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-config-data\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.516146 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-scripts\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.516422 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-config-data\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.525465 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4bf21836-6124-4093-89ce-5512a9239d73-horizon-secret-key\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.532207 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-db-sync-config-data\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.532684 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-combined-ca-bundle\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.548152 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.568554 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.575310 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.578639 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.578834 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.579212 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ngd6\" (UniqueName: \"kubernetes.io/projected/4bf21836-6124-4093-89ce-5512a9239d73-kube-api-access-8ngd6\") pod \"horizon-5648749bd9-8rckr\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.585478 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.598663 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-run-httpd\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.598812 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-config-data\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.599013 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-log-httpd\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.599087 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.599116 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-scripts\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.599152 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.599217 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmx6l\" (UniqueName: \"kubernetes.io/projected/92bcbff2-f0a3-4303-96cc-5f3177731f6a-kube-api-access-qmx6l\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.624949 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qggnm\" (UniqueName: \"kubernetes.io/projected/e199d2a8-2a58-47c7-8018-13b29c37c2ad-kube-api-access-qggnm\") pod \"cinder-db-sync-z7k2l\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.627249 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.643363 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.660114 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-688b69fc-7dd7f"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.662010 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.670143 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b74bb6999-gs5cq"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.677806 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-k5dlb"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.678940 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.681851 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-bfjm2" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.682133 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.703736 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-scripts\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.703784 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-log-httpd\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.703846 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.703877 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-scripts\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.703916 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.703977 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmx6l\" (UniqueName: \"kubernetes.io/projected/92bcbff2-f0a3-4303-96cc-5f3177731f6a-kube-api-access-qmx6l\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.704032 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb55042a-f0de-40dd-8a25-49a0226f4a8b-horizon-secret-key\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.704079 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-run-httpd\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.704099 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-config-data\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.704163 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb55042a-f0de-40dd-8a25-49a0226f4a8b-logs\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.704210 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55dcn\" (UniqueName: \"kubernetes.io/projected/cb55042a-f0de-40dd-8a25-49a0226f4a8b-kube-api-access-55dcn\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.704234 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-config-data\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.704986 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-log-httpd\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.706640 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-run-httpd\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.708848 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-scripts\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.708896 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-688b69fc-7dd7f"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.712866 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.714953 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-config-data\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.715181 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-d2fzh"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.725673 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.737197 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.738718 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-2vm9f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.739485 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.740393 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.740889 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-k5dlb"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.763862 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-d2fzh"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.783610 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmx6l\" (UniqueName: \"kubernetes.io/projected/92bcbff2-f0a3-4303-96cc-5f3177731f6a-kube-api-access-qmx6l\") pod \"ceilometer-0\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.787331 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-8xkhj"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.790705 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.794495 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.794802 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-szzm6" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.795606 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.807436 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-scripts\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.807522 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-combined-ca-bundle\") pod \"barbican-db-sync-k5dlb\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.807566 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-db-sync-config-data\") pod \"barbican-db-sync-k5dlb\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.807614 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb55042a-f0de-40dd-8a25-49a0226f4a8b-horizon-secret-key\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.807640 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmqbx\" (UniqueName: \"kubernetes.io/projected/007f0b25-94b5-4529-b835-08cdc34e0d6e-kube-api-access-qmqbx\") pod \"neutron-db-sync-d2fzh\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.807674 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-config-data\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.807696 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-combined-ca-bundle\") pod \"neutron-db-sync-d2fzh\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.816523 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb55042a-f0de-40dd-8a25-49a0226f4a8b-logs\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.816615 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55dcn\" (UniqueName: \"kubernetes.io/projected/cb55042a-f0de-40dd-8a25-49a0226f4a8b-kube-api-access-55dcn\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.816705 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdl48\" (UniqueName: \"kubernetes.io/projected/913aa7a2-f310-4b1e-a403-2e5f98b339b5-kube-api-access-fdl48\") pod \"barbican-db-sync-k5dlb\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.816785 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-config\") pod \"neutron-db-sync-d2fzh\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.823762 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb55042a-f0de-40dd-8a25-49a0226f4a8b-logs\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.825060 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-scripts\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.826348 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-config-data\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.828342 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb55042a-f0de-40dd-8a25-49a0226f4a8b-horizon-secret-key\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.830256 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-548f9979d5-zkcjj"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.842027 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-548f9979d5-zkcjj"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.842147 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.847451 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8xkhj"] Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.872448 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55dcn\" (UniqueName: \"kubernetes.io/projected/cb55042a-f0de-40dd-8a25-49a0226f4a8b-kube-api-access-55dcn\") pod \"horizon-688b69fc-7dd7f\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.925149 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.930352 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-svc\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.930423 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdl48\" (UniqueName: \"kubernetes.io/projected/913aa7a2-f310-4b1e-a403-2e5f98b339b5-kube-api-access-fdl48\") pod \"barbican-db-sync-k5dlb\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.930457 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-config\") pod \"neutron-db-sync-d2fzh\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.930571 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-combined-ca-bundle\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.930661 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-config\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.935872 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-combined-ca-bundle\") pod \"barbican-db-sync-k5dlb\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.935971 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-scripts\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.936030 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-db-sync-config-data\") pod \"barbican-db-sync-k5dlb\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.936141 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmqbx\" (UniqueName: \"kubernetes.io/projected/007f0b25-94b5-4529-b835-08cdc34e0d6e-kube-api-access-qmqbx\") pod \"neutron-db-sync-d2fzh\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.936174 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2h6n\" (UniqueName: \"kubernetes.io/projected/0f887e1a-55da-42cd-a8f1-709f797b283e-kube-api-access-r2h6n\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.936208 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-swift-storage-0\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.936259 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-combined-ca-bundle\") pod \"neutron-db-sync-d2fzh\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.936301 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-logs\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.936327 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-config-data\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.936414 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-nb\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.936444 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-sb\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.936492 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqkqq\" (UniqueName: \"kubernetes.io/projected/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-kube-api-access-bqkqq\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.940656 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-combined-ca-bundle\") pod \"barbican-db-sync-k5dlb\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.942319 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-db-sync-config-data\") pod \"barbican-db-sync-k5dlb\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.944426 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-config\") pod \"neutron-db-sync-d2fzh\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.949993 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-combined-ca-bundle\") pod \"neutron-db-sync-d2fzh\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.950217 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdl48\" (UniqueName: \"kubernetes.io/projected/913aa7a2-f310-4b1e-a403-2e5f98b339b5-kube-api-access-fdl48\") pod \"barbican-db-sync-k5dlb\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.953777 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmqbx\" (UniqueName: \"kubernetes.io/projected/007f0b25-94b5-4529-b835-08cdc34e0d6e-kube-api-access-qmqbx\") pod \"neutron-db-sync-d2fzh\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.967824 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:21:43 crc kubenswrapper[4810]: I0930 08:21:43.983666 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.002718 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.040958 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-nb\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.041000 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-sb\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.041034 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqkqq\" (UniqueName: \"kubernetes.io/projected/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-kube-api-access-bqkqq\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.041068 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-svc\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.041102 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-combined-ca-bundle\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.041142 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-config\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.041166 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-scripts\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.041214 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2h6n\" (UniqueName: \"kubernetes.io/projected/0f887e1a-55da-42cd-a8f1-709f797b283e-kube-api-access-r2h6n\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.041235 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-swift-storage-0\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.041259 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-logs\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.041292 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-config-data\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.042809 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-config\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.043951 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-sb\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.044758 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-nb\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.045697 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-logs\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.045838 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-config-data\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.046340 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-swift-storage-0\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.046923 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-svc\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.051016 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-combined-ca-bundle\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.054520 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-scripts\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.060944 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.063439 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqkqq\" (UniqueName: \"kubernetes.io/projected/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-kube-api-access-bqkqq\") pod \"placement-db-sync-8xkhj\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.066876 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2h6n\" (UniqueName: \"kubernetes.io/projected/0f887e1a-55da-42cd-a8f1-709f797b283e-kube-api-access-r2h6n\") pod \"dnsmasq-dns-548f9979d5-zkcjj\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.072525 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8xbnb"] Sep 30 08:21:44 crc kubenswrapper[4810]: W0930 08:21:44.081756 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ed92924_5cd6_4d38_ac53_7db3c0773437.slice/crio-fb558d3834535a88bde0d6f5d7abdc98a8f1e52f4f79dd5d8b20fc58d28333c7 WatchSource:0}: Error finding container fb558d3834535a88bde0d6f5d7abdc98a8f1e52f4f79dd5d8b20fc58d28333c7: Status 404 returned error can't find the container with id fb558d3834535a88bde0d6f5d7abdc98a8f1e52f4f79dd5d8b20fc58d28333c7 Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.139580 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8xkhj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.177743 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.260461 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b74bb6999-gs5cq"] Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.521697 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8xbnb" event={"ID":"8ed92924-5cd6-4d38-ac53-7db3c0773437","Type":"ContainerStarted","Data":"fb558d3834535a88bde0d6f5d7abdc98a8f1e52f4f79dd5d8b20fc58d28333c7"} Sep 30 08:21:44 crc kubenswrapper[4810]: I0930 08:21:44.523379 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" event={"ID":"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e","Type":"ContainerStarted","Data":"50e0d3c0b879b7bc55538c6a77f1c3eefa80898d95e407679eefcef305084146"} Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.466317 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-k5dlb"] Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.476759 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-d2fzh"] Sep 30 08:21:45 crc kubenswrapper[4810]: W0930 08:21:45.482901 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod007f0b25_94b5_4529_b835_08cdc34e0d6e.slice/crio-17407992b4fc4da5bda22c275094a06296d27030eb09a396f535b6414bba7f18 WatchSource:0}: Error finding container 17407992b4fc4da5bda22c275094a06296d27030eb09a396f535b6414bba7f18: Status 404 returned error can't find the container with id 17407992b4fc4da5bda22c275094a06296d27030eb09a396f535b6414bba7f18 Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.504321 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-z7k2l"] Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.519000 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.526575 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.536878 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-688b69fc-7dd7f"] Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.557583 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.560858 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k5dlb" event={"ID":"913aa7a2-f310-4b1e-a403-2e5f98b339b5","Type":"ContainerStarted","Data":"edcc2f23746b8a250d9181834cc31539888489f4a716c9584e4b00ed6860b8d0"} Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.564759 4810 generic.go:334] "Generic (PLEG): container finished" podID="a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e" containerID="f22ec4f67d23b1a72ffb1e387156b5bdbf4c111d66a36b3739f7821bc21c4c94" exitCode=0 Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.564808 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" event={"ID":"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e","Type":"ContainerDied","Data":"f22ec4f67d23b1a72ffb1e387156b5bdbf4c111d66a36b3739f7821bc21c4c94"} Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.570456 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-z7k2l" event={"ID":"e199d2a8-2a58-47c7-8018-13b29c37c2ad","Type":"ContainerStarted","Data":"54d060a5e6c7bd24bf9c16899a43f288f357b52be1955c22fec188188163b352"} Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.627520 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d2fzh" event={"ID":"007f0b25-94b5-4529-b835-08cdc34e0d6e","Type":"ContainerStarted","Data":"17407992b4fc4da5bda22c275094a06296d27030eb09a396f535b6414bba7f18"} Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.631308 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerStarted","Data":"00a6fd5584e4f390a6774bbf247a0d5d9b5bac7c959fd97b7fce55ccefbd375d"} Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.635143 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8xbnb" event={"ID":"8ed92924-5cd6-4d38-ac53-7db3c0773437","Type":"ContainerStarted","Data":"e3e793bcf7de6abdc5f73e4768fbecd1d417ac36ed1653e5ac5026c8ac59b493"} Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.639324 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-688b69fc-7dd7f" event={"ID":"cb55042a-f0de-40dd-8a25-49a0226f4a8b","Type":"ContainerStarted","Data":"77bd1b65e7ce45798341795fc0d2d38ad2baa54ee2d5a3c044f3acc694caa2fd"} Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.720719 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-8xbnb" podStartSLOduration=3.720696193 podStartE2EDuration="3.720696193s" podCreationTimestamp="2025-09-30 08:21:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:21:45.665475729 +0000 UTC m=+1129.117674996" watchObservedRunningTime="2025-09-30 08:21:45.720696193 +0000 UTC m=+1129.172895460" Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.816381 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5648749bd9-8rckr"] Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.838175 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-98d79fd7c-kz5wm"] Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.857964 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.902955 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-98d79fd7c-kz5wm"] Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.936051 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.975778 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:21:45 crc kubenswrapper[4810]: I0930 08:21:45.996098 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5648749bd9-8rckr"] Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.013158 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-logs\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.013468 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw47h\" (UniqueName: \"kubernetes.io/projected/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-kube-api-access-hw47h\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.013494 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-config-data\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.013642 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-horizon-secret-key\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.013687 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-scripts\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.025745 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-548f9979d5-zkcjj"] Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.042567 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.073725 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8xkhj"] Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.118327 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-logs\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.118396 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw47h\" (UniqueName: \"kubernetes.io/projected/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-kube-api-access-hw47h\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.118430 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-config-data\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.118486 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-horizon-secret-key\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.118523 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-scripts\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.119284 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-logs\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.119315 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-scripts\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.127571 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-horizon-secret-key\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.131309 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-config-data\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.161968 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw47h\" (UniqueName: \"kubernetes.io/projected/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-kube-api-access-hw47h\") pod \"horizon-98d79fd7c-kz5wm\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: E0930 08:21:46.184818 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dacf67c_34b5_4c58_9f8d_00adfdcf74b7.slice\": RecentStats: unable to find data in memory cache]" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.212680 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.322525 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-config\") pod \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.322976 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-sb\") pod \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.324866 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-swift-storage-0\") pod \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.324935 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-nb\") pod \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.325087 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-svc\") pod \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.325122 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbfnt\" (UniqueName: \"kubernetes.io/projected/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-kube-api-access-xbfnt\") pod \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\" (UID: \"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e\") " Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.328747 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.355220 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e" (UID: "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.356658 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e" (UID: "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.356810 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-kube-api-access-xbfnt" (OuterVolumeSpecName: "kube-api-access-xbfnt") pod "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e" (UID: "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e"). InnerVolumeSpecName "kube-api-access-xbfnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.360976 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-config" (OuterVolumeSpecName: "config") pod "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e" (UID: "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.368597 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e" (UID: "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.379927 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e" (UID: "a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.428345 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.428387 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.428398 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.428407 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.428415 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.428424 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbfnt\" (UniqueName: \"kubernetes.io/projected/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e-kube-api-access-xbfnt\") on node \"crc\" DevicePath \"\"" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.670816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92bcbff2-f0a3-4303-96cc-5f3177731f6a","Type":"ContainerStarted","Data":"13b542e536c8600b9a6c7206413ee3fc3f8b1032c6623429f867b1108a6de950"} Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.681058 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.683445 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b74bb6999-gs5cq" event={"ID":"a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e","Type":"ContainerDied","Data":"50e0d3c0b879b7bc55538c6a77f1c3eefa80898d95e407679eefcef305084146"} Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.683545 4810 scope.go:117] "RemoveContainer" containerID="f22ec4f67d23b1a72ffb1e387156b5bdbf4c111d66a36b3739f7821bc21c4c94" Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.750801 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b74bb6999-gs5cq"] Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.754252 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bc876eb0-0451-496a-90c1-4470a9d21ca2","Type":"ContainerStarted","Data":"cb5133677659ac907189465c3258647ef90e5b32193cd0cf7ce8fb0f5d05ea64"} Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.754308 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bc876eb0-0451-496a-90c1-4470a9d21ca2","Type":"ContainerStarted","Data":"b297c8d27efcc4b8c686bb4b2ccd3a2f87bc0cb24596e57aa1bab0fe826287fe"} Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.762455 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b74bb6999-gs5cq"] Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.764987 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d2fzh" event={"ID":"007f0b25-94b5-4529-b835-08cdc34e0d6e","Type":"ContainerStarted","Data":"96a88818d3c567e4575fc2096f67461bfab333bbf939f7fa14727864172c5f4f"} Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.769433 4810 generic.go:334] "Generic (PLEG): container finished" podID="0f887e1a-55da-42cd-a8f1-709f797b283e" containerID="e52ada73ebee88d08bce03eb187c0069f0d9fd0c6a24f068dc26fd3191ace16b" exitCode=0 Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.769501 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" event={"ID":"0f887e1a-55da-42cd-a8f1-709f797b283e","Type":"ContainerDied","Data":"e52ada73ebee88d08bce03eb187c0069f0d9fd0c6a24f068dc26fd3191ace16b"} Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.769529 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" event={"ID":"0f887e1a-55da-42cd-a8f1-709f797b283e","Type":"ContainerStarted","Data":"3f84a89dda6951f78d490d5b349989c1beba322fae674b13b9f7661976126240"} Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.779708 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"277a70ee-91dd-4198-9abd-7e6bb909bec0","Type":"ContainerStarted","Data":"4f4094cc21166c6514efd67967a1643035a9b7d43b18dfee4882fcd89e6c90b6"} Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.789632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8xkhj" event={"ID":"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f","Type":"ContainerStarted","Data":"5b8a148dc3a7f0c593194596567d8de523c4b7dbb0f85955593ad2971bd011a7"} Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.792384 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5648749bd9-8rckr" event={"ID":"4bf21836-6124-4093-89ce-5512a9239d73","Type":"ContainerStarted","Data":"e202cb1e140f5f7a0337b544bea017b26339cf6186d267ce83427c97ecdb317c"} Sep 30 08:21:46 crc kubenswrapper[4810]: I0930 08:21:46.867476 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-d2fzh" podStartSLOduration=3.867453034 podStartE2EDuration="3.867453034s" podCreationTimestamp="2025-09-30 08:21:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:21:46.80399746 +0000 UTC m=+1130.256196727" watchObservedRunningTime="2025-09-30 08:21:46.867453034 +0000 UTC m=+1130.319652301" Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.162196 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-98d79fd7c-kz5wm"] Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.342822 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e" path="/var/lib/kubelet/pods/a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e/volumes" Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.831902 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" event={"ID":"0f887e1a-55da-42cd-a8f1-709f797b283e","Type":"ContainerStarted","Data":"6a26e395aab71c0f229066ce5c97280e9ba8cb043f774c09c05d83df7239c0b4"} Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.833022 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.843593 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-98d79fd7c-kz5wm" event={"ID":"dc2cb8f1-4585-43b1-865c-bea9b61ef20c","Type":"ContainerStarted","Data":"35e373788398a18fee8bdf3c985ae464a5452dca2be0c15ccca9ae00f76b7502"} Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.861467 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" podStartSLOduration=4.861442866 podStartE2EDuration="4.861442866s" podCreationTimestamp="2025-09-30 08:21:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:21:47.855793442 +0000 UTC m=+1131.307992709" watchObservedRunningTime="2025-09-30 08:21:47.861442866 +0000 UTC m=+1131.313642143" Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.867280 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api-log" containerID="cri-o://cb5133677659ac907189465c3258647ef90e5b32193cd0cf7ce8fb0f5d05ea64" gracePeriod=30 Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.867389 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bc876eb0-0451-496a-90c1-4470a9d21ca2","Type":"ContainerStarted","Data":"a760f51a5ae86a3ff90e25446866e430fef5f31536ab54901f0472747b3f8418"} Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.867659 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api" containerID="cri-o://a760f51a5ae86a3ff90e25446866e430fef5f31536ab54901f0472747b3f8418" gracePeriod=30 Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.868060 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.876453 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.153:9322/\": EOF" Sep 30 08:21:47 crc kubenswrapper[4810]: I0930 08:21:47.891741 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=4.891721436 podStartE2EDuration="4.891721436s" podCreationTimestamp="2025-09-30 08:21:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:21:47.88187775 +0000 UTC m=+1131.334077017" watchObservedRunningTime="2025-09-30 08:21:47.891721436 +0000 UTC m=+1131.343920703" Sep 30 08:21:48 crc kubenswrapper[4810]: I0930 08:21:48.549108 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Sep 30 08:21:48 crc kubenswrapper[4810]: I0930 08:21:48.903508 4810 generic.go:334] "Generic (PLEG): container finished" podID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerID="cb5133677659ac907189465c3258647ef90e5b32193cd0cf7ce8fb0f5d05ea64" exitCode=143 Sep 30 08:21:48 crc kubenswrapper[4810]: I0930 08:21:48.903585 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bc876eb0-0451-496a-90c1-4470a9d21ca2","Type":"ContainerDied","Data":"cb5133677659ac907189465c3258647ef90e5b32193cd0cf7ce8fb0f5d05ea64"} Sep 30 08:21:51 crc kubenswrapper[4810]: I0930 08:21:51.616449 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.153:9322/\": read tcp 10.217.0.2:57612->10.217.0.153:9322: read: connection reset by peer" Sep 30 08:21:51 crc kubenswrapper[4810]: I0930 08:21:51.931597 4810 generic.go:334] "Generic (PLEG): container finished" podID="8ed92924-5cd6-4d38-ac53-7db3c0773437" containerID="e3e793bcf7de6abdc5f73e4768fbecd1d417ac36ed1653e5ac5026c8ac59b493" exitCode=0 Sep 30 08:21:51 crc kubenswrapper[4810]: I0930 08:21:51.931686 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8xbnb" event={"ID":"8ed92924-5cd6-4d38-ac53-7db3c0773437","Type":"ContainerDied","Data":"e3e793bcf7de6abdc5f73e4768fbecd1d417ac36ed1653e5ac5026c8ac59b493"} Sep 30 08:21:51 crc kubenswrapper[4810]: I0930 08:21:51.946228 4810 generic.go:334] "Generic (PLEG): container finished" podID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerID="a760f51a5ae86a3ff90e25446866e430fef5f31536ab54901f0472747b3f8418" exitCode=0 Sep 30 08:21:51 crc kubenswrapper[4810]: I0930 08:21:51.946282 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bc876eb0-0451-496a-90c1-4470a9d21ca2","Type":"ContainerDied","Data":"a760f51a5ae86a3ff90e25446866e430fef5f31536ab54901f0472747b3f8418"} Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.398306 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-688b69fc-7dd7f"] Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.437645 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-754cbc7dbb-wbdt6"] Sep 30 08:21:52 crc kubenswrapper[4810]: E0930 08:21:52.438142 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e" containerName="init" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.438153 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e" containerName="init" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.438365 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ea61f0-fbe1-4fbc-a073-749fc1c1d72e" containerName="init" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.439367 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.445214 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.466215 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-754cbc7dbb-wbdt6"] Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.495248 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-98d79fd7c-kz5wm"] Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.512209 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d769bd5f4-zkxj4"] Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.514262 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.529124 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d769bd5f4-zkxj4"] Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.598494 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-secret-key\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.598619 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-combined-ca-bundle\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.598652 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-scripts\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.598677 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6tx2\" (UniqueName: \"kubernetes.io/projected/c176d78f-1abf-4191-94f8-8ee7be102b18-kube-api-access-q6tx2\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.598720 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-tls-certs\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.598752 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-config-data\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.598772 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c176d78f-1abf-4191-94f8-8ee7be102b18-logs\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702194 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8fwc\" (UniqueName: \"kubernetes.io/projected/a71102a5-388b-4dcb-bd80-db92ed28f1f5-kube-api-access-m8fwc\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702306 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-combined-ca-bundle\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702366 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-scripts\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702387 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a71102a5-388b-4dcb-bd80-db92ed28f1f5-scripts\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702408 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6tx2\" (UniqueName: \"kubernetes.io/projected/c176d78f-1abf-4191-94f8-8ee7be102b18-kube-api-access-q6tx2\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702424 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-tls-certs\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702440 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a71102a5-388b-4dcb-bd80-db92ed28f1f5-horizon-secret-key\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702461 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a71102a5-388b-4dcb-bd80-db92ed28f1f5-config-data\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702484 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a71102a5-388b-4dcb-bd80-db92ed28f1f5-combined-ca-bundle\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702501 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-config-data\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702517 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a71102a5-388b-4dcb-bd80-db92ed28f1f5-logs\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702536 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c176d78f-1abf-4191-94f8-8ee7be102b18-logs\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-secret-key\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.702910 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a71102a5-388b-4dcb-bd80-db92ed28f1f5-horizon-tls-certs\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.704100 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c176d78f-1abf-4191-94f8-8ee7be102b18-logs\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.704371 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-config-data\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.704444 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-scripts\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.708851 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-tls-certs\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.713930 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-combined-ca-bundle\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.737874 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-secret-key\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.738147 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6tx2\" (UniqueName: \"kubernetes.io/projected/c176d78f-1abf-4191-94f8-8ee7be102b18-kube-api-access-q6tx2\") pod \"horizon-754cbc7dbb-wbdt6\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.771324 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.805238 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a71102a5-388b-4dcb-bd80-db92ed28f1f5-scripts\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.805301 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a71102a5-388b-4dcb-bd80-db92ed28f1f5-horizon-secret-key\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.805321 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a71102a5-388b-4dcb-bd80-db92ed28f1f5-config-data\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.805341 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a71102a5-388b-4dcb-bd80-db92ed28f1f5-combined-ca-bundle\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.805357 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a71102a5-388b-4dcb-bd80-db92ed28f1f5-logs\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.805446 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a71102a5-388b-4dcb-bd80-db92ed28f1f5-horizon-tls-certs\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.805521 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8fwc\" (UniqueName: \"kubernetes.io/projected/a71102a5-388b-4dcb-bd80-db92ed28f1f5-kube-api-access-m8fwc\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.807088 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a71102a5-388b-4dcb-bd80-db92ed28f1f5-scripts\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.831473 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a71102a5-388b-4dcb-bd80-db92ed28f1f5-horizon-secret-key\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.832318 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a71102a5-388b-4dcb-bd80-db92ed28f1f5-logs\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.832779 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a71102a5-388b-4dcb-bd80-db92ed28f1f5-config-data\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.839397 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a71102a5-388b-4dcb-bd80-db92ed28f1f5-combined-ca-bundle\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.839410 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a71102a5-388b-4dcb-bd80-db92ed28f1f5-horizon-tls-certs\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:52 crc kubenswrapper[4810]: I0930 08:21:52.857663 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8fwc\" (UniqueName: \"kubernetes.io/projected/a71102a5-388b-4dcb-bd80-db92ed28f1f5-kube-api-access-m8fwc\") pod \"horizon-6d769bd5f4-zkxj4\" (UID: \"a71102a5-388b-4dcb-bd80-db92ed28f1f5\") " pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:53 crc kubenswrapper[4810]: I0930 08:21:53.144873 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:21:54 crc kubenswrapper[4810]: I0930 08:21:54.179896 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:21:54 crc kubenswrapper[4810]: I0930 08:21:54.280751 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b9b87fb55-7jxzz"] Sep 30 08:21:54 crc kubenswrapper[4810]: I0930 08:21:54.281057 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" podUID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerName="dnsmasq-dns" containerID="cri-o://11c46bda01fa7f7f62f2e761f2232d7700b63962f2aa93f556d70bd51276e6dd" gracePeriod=10 Sep 30 08:21:54 crc kubenswrapper[4810]: I0930 08:21:54.990068 4810 generic.go:334] "Generic (PLEG): container finished" podID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerID="11c46bda01fa7f7f62f2e761f2232d7700b63962f2aa93f556d70bd51276e6dd" exitCode=0 Sep 30 08:21:54 crc kubenswrapper[4810]: I0930 08:21:54.990126 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" event={"ID":"cf25d291-3d0d-48c0-9dae-405ef9fd4922","Type":"ContainerDied","Data":"11c46bda01fa7f7f62f2e761f2232d7700b63962f2aa93f556d70bd51276e6dd"} Sep 30 08:21:56 crc kubenswrapper[4810]: E0930 08:21:56.441819 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dacf67c_34b5_4c58_9f8d_00adfdcf74b7.slice\": RecentStats: unable to find data in memory cache]" Sep 30 08:21:56 crc kubenswrapper[4810]: I0930 08:21:56.562989 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" podUID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: connect: connection refused" Sep 30 08:21:58 crc kubenswrapper[4810]: I0930 08:21:58.548743 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.153:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 08:22:01 crc kubenswrapper[4810]: I0930 08:22:01.562792 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" podUID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: connect: connection refused" Sep 30 08:22:03 crc kubenswrapper[4810]: I0930 08:22:03.550094 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.153:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 08:22:03 crc kubenswrapper[4810]: E0930 08:22:03.814163 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-horizon:watcher_latest" Sep 30 08:22:03 crc kubenswrapper[4810]: E0930 08:22:03.814245 4810 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-horizon:watcher_latest" Sep 30 08:22:03 crc kubenswrapper[4810]: E0930 08:22:03.814575 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.41:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n676hf7h5cfhf8h86hfch65fh55ch575h58fh59bh64dh569h9dh647h5f5h5bbh5c8h6h556h5dh547h65ch578h96h66chf6h675h5fh68h569h5dbq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55dcn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-688b69fc-7dd7f_openstack(cb55042a-f0de-40dd-8a25-49a0226f4a8b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 08:22:03 crc kubenswrapper[4810]: E0930 08:22:03.817444 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.41:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-688b69fc-7dd7f" podUID="cb55042a-f0de-40dd-8a25-49a0226f4a8b" Sep 30 08:22:04 crc kubenswrapper[4810]: E0930 08:22:04.189965 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Sep 30 08:22:04 crc kubenswrapper[4810]: E0930 08:22:04.190013 4810 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Sep 30 08:22:04 crc kubenswrapper[4810]: E0930 08:22:04.190116 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:38.102.83.41:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n57dh84hd5h65dhdh675h54h69h5ddh5f8h598h5f7h59fh68fh8bh579h6h578h5b7h57fh9h55bh8dh55fh57h5cch5fbh79h6fh594h64hbq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qmx6l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(92bcbff2-f0a3-4303-96cc-5f3177731f6a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.196055 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 08:22:04 crc kubenswrapper[4810]: E0930 08:22:04.223027 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-horizon:watcher_latest" Sep 30 08:22:04 crc kubenswrapper[4810]: E0930 08:22:04.223352 4810 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-horizon:watcher_latest" Sep 30 08:22:04 crc kubenswrapper[4810]: E0930 08:22:04.223501 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.41:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n55chc9h84h67ch569hfbhb4hddhcbh55fh58fh5b9h5d4h565h5bh574hb6hd7h554h8dh64ch5b5h7fh674h5fh76h78h86h56fh59h699h587q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8ngd6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5648749bd9-8rckr_openstack(4bf21836-6124-4093-89ce-5512a9239d73): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 08:22:04 crc kubenswrapper[4810]: E0930 08:22:04.226397 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.41:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-5648749bd9-8rckr" podUID="4bf21836-6124-4093-89ce-5512a9239d73" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.353344 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.358639 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.457021 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-combined-ca-bundle\") pod \"bc876eb0-0451-496a-90c1-4470a9d21ca2\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.457082 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-combined-ca-bundle\") pod \"8ed92924-5cd6-4d38-ac53-7db3c0773437\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.457113 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dkwf\" (UniqueName: \"kubernetes.io/projected/bc876eb0-0451-496a-90c1-4470a9d21ca2-kube-api-access-5dkwf\") pod \"bc876eb0-0451-496a-90c1-4470a9d21ca2\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.457153 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sml94\" (UniqueName: \"kubernetes.io/projected/8ed92924-5cd6-4d38-ac53-7db3c0773437-kube-api-access-sml94\") pod \"8ed92924-5cd6-4d38-ac53-7db3c0773437\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.457180 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-config-data\") pod \"8ed92924-5cd6-4d38-ac53-7db3c0773437\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.457328 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-custom-prometheus-ca\") pod \"bc876eb0-0451-496a-90c1-4470a9d21ca2\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.457453 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-scripts\") pod \"8ed92924-5cd6-4d38-ac53-7db3c0773437\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.457481 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc876eb0-0451-496a-90c1-4470a9d21ca2-logs\") pod \"bc876eb0-0451-496a-90c1-4470a9d21ca2\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.457526 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-credential-keys\") pod \"8ed92924-5cd6-4d38-ac53-7db3c0773437\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.457561 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-config-data\") pod \"bc876eb0-0451-496a-90c1-4470a9d21ca2\" (UID: \"bc876eb0-0451-496a-90c1-4470a9d21ca2\") " Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.457583 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-fernet-keys\") pod \"8ed92924-5cd6-4d38-ac53-7db3c0773437\" (UID: \"8ed92924-5cd6-4d38-ac53-7db3c0773437\") " Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.458020 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc876eb0-0451-496a-90c1-4470a9d21ca2-logs" (OuterVolumeSpecName: "logs") pod "bc876eb0-0451-496a-90c1-4470a9d21ca2" (UID: "bc876eb0-0451-496a-90c1-4470a9d21ca2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.458966 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc876eb0-0451-496a-90c1-4470a9d21ca2-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.479650 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8ed92924-5cd6-4d38-ac53-7db3c0773437" (UID: "8ed92924-5cd6-4d38-ac53-7db3c0773437"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.479890 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8ed92924-5cd6-4d38-ac53-7db3c0773437" (UID: "8ed92924-5cd6-4d38-ac53-7db3c0773437"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.480203 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ed92924-5cd6-4d38-ac53-7db3c0773437-kube-api-access-sml94" (OuterVolumeSpecName: "kube-api-access-sml94") pod "8ed92924-5cd6-4d38-ac53-7db3c0773437" (UID: "8ed92924-5cd6-4d38-ac53-7db3c0773437"). InnerVolumeSpecName "kube-api-access-sml94". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.481340 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-scripts" (OuterVolumeSpecName: "scripts") pod "8ed92924-5cd6-4d38-ac53-7db3c0773437" (UID: "8ed92924-5cd6-4d38-ac53-7db3c0773437"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.499926 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc876eb0-0451-496a-90c1-4470a9d21ca2-kube-api-access-5dkwf" (OuterVolumeSpecName: "kube-api-access-5dkwf") pod "bc876eb0-0451-496a-90c1-4470a9d21ca2" (UID: "bc876eb0-0451-496a-90c1-4470a9d21ca2"). InnerVolumeSpecName "kube-api-access-5dkwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.501205 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "bc876eb0-0451-496a-90c1-4470a9d21ca2" (UID: "bc876eb0-0451-496a-90c1-4470a9d21ca2"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.503961 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ed92924-5cd6-4d38-ac53-7db3c0773437" (UID: "8ed92924-5cd6-4d38-ac53-7db3c0773437"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.519920 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc876eb0-0451-496a-90c1-4470a9d21ca2" (UID: "bc876eb0-0451-496a-90c1-4470a9d21ca2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.521826 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-config-data" (OuterVolumeSpecName: "config-data") pod "8ed92924-5cd6-4d38-ac53-7db3c0773437" (UID: "8ed92924-5cd6-4d38-ac53-7db3c0773437"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.536169 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-config-data" (OuterVolumeSpecName: "config-data") pod "bc876eb0-0451-496a-90c1-4470a9d21ca2" (UID: "bc876eb0-0451-496a-90c1-4470a9d21ca2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.564220 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sml94\" (UniqueName: \"kubernetes.io/projected/8ed92924-5cd6-4d38-ac53-7db3c0773437-kube-api-access-sml94\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.564282 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.564297 4810 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.564307 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.564319 4810 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.564329 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.564338 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.564348 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc876eb0-0451-496a-90c1-4470a9d21ca2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.564358 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed92924-5cd6-4d38-ac53-7db3c0773437-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:04 crc kubenswrapper[4810]: I0930 08:22:04.564367 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dkwf\" (UniqueName: \"kubernetes.io/projected/bc876eb0-0451-496a-90c1-4470a9d21ca2-kube-api-access-5dkwf\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.163686 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8xbnb" event={"ID":"8ed92924-5cd6-4d38-ac53-7db3c0773437","Type":"ContainerDied","Data":"fb558d3834535a88bde0d6f5d7abdc98a8f1e52f4f79dd5d8b20fc58d28333c7"} Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.163745 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb558d3834535a88bde0d6f5d7abdc98a8f1e52f4f79dd5d8b20fc58d28333c7" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.163827 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8xbnb" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.175418 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bc876eb0-0451-496a-90c1-4470a9d21ca2","Type":"ContainerDied","Data":"b297c8d27efcc4b8c686bb4b2ccd3a2f87bc0cb24596e57aa1bab0fe826287fe"} Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.175633 4810 scope.go:117] "RemoveContainer" containerID="a760f51a5ae86a3ff90e25446866e430fef5f31536ab54901f0472747b3f8418" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.175450 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.257792 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.265736 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.284102 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:22:05 crc kubenswrapper[4810]: E0930 08:22:05.284473 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api-log" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.284488 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api-log" Sep 30 08:22:05 crc kubenswrapper[4810]: E0930 08:22:05.284504 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed92924-5cd6-4d38-ac53-7db3c0773437" containerName="keystone-bootstrap" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.284511 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed92924-5cd6-4d38-ac53-7db3c0773437" containerName="keystone-bootstrap" Sep 30 08:22:05 crc kubenswrapper[4810]: E0930 08:22:05.284530 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.284537 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.284715 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api-log" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.284736 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ed92924-5cd6-4d38-ac53-7db3c0773437" containerName="keystone-bootstrap" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.284746 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.285655 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.290684 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.335529 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" path="/var/lib/kubelet/pods/bc876eb0-0451-496a-90c1-4470a9d21ca2/volumes" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.336157 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.379493 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acf21e98-01dc-4a20-aa79-bb5549e59f22-logs\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.379745 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.379799 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-config-data\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.379989 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.380029 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snxfp\" (UniqueName: \"kubernetes.io/projected/acf21e98-01dc-4a20-aa79-bb5549e59f22-kube-api-access-snxfp\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.481404 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acf21e98-01dc-4a20-aa79-bb5549e59f22-logs\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.481694 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.481799 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-config-data\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.482007 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.482084 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snxfp\" (UniqueName: \"kubernetes.io/projected/acf21e98-01dc-4a20-aa79-bb5549e59f22-kube-api-access-snxfp\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.482806 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acf21e98-01dc-4a20-aa79-bb5549e59f22-logs\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.487781 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.489005 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-config-data\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.496420 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.497861 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snxfp\" (UniqueName: \"kubernetes.io/projected/acf21e98-01dc-4a20-aa79-bb5549e59f22-kube-api-access-snxfp\") pod \"watcher-api-0\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.551985 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-8xbnb"] Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.561029 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-8xbnb"] Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.606398 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.652498 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-x57vl"] Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.653975 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.656485 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.656719 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-hw6qv" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.657858 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.657987 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.662449 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-x57vl"] Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.689185 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t5pn\" (UniqueName: \"kubernetes.io/projected/65f3abf6-0015-41fc-9a40-0ed346bb84ec-kube-api-access-4t5pn\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.689247 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-credential-keys\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.689315 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-fernet-keys\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.689359 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-scripts\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.689379 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-config-data\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.689412 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-combined-ca-bundle\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.790245 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t5pn\" (UniqueName: \"kubernetes.io/projected/65f3abf6-0015-41fc-9a40-0ed346bb84ec-kube-api-access-4t5pn\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.790320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-credential-keys\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.790347 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-fernet-keys\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.790388 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-scripts\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.790405 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-config-data\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.790439 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-combined-ca-bundle\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.795206 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-config-data\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.795921 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-combined-ca-bundle\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.796215 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-fernet-keys\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.796550 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-credential-keys\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.796766 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-scripts\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.816406 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t5pn\" (UniqueName: \"kubernetes.io/projected/65f3abf6-0015-41fc-9a40-0ed346bb84ec-kube-api-access-4t5pn\") pod \"keystone-bootstrap-x57vl\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:05 crc kubenswrapper[4810]: I0930 08:22:05.977561 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:07 crc kubenswrapper[4810]: I0930 08:22:07.325055 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ed92924-5cd6-4d38-ac53-7db3c0773437" path="/var/lib/kubelet/pods/8ed92924-5cd6-4d38-ac53-7db3c0773437/volumes" Sep 30 08:22:08 crc kubenswrapper[4810]: I0930 08:22:08.551489 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="bc876eb0-0451-496a-90c1-4470a9d21ca2" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.153:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 08:22:11 crc kubenswrapper[4810]: I0930 08:22:11.562550 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" podUID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: i/o timeout" Sep 30 08:22:11 crc kubenswrapper[4810]: I0930 08:22:11.562952 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:22:12 crc kubenswrapper[4810]: I0930 08:22:12.246463 4810 generic.go:334] "Generic (PLEG): container finished" podID="007f0b25-94b5-4529-b835-08cdc34e0d6e" containerID="96a88818d3c567e4575fc2096f67461bfab333bbf939f7fa14727864172c5f4f" exitCode=0 Sep 30 08:22:12 crc kubenswrapper[4810]: I0930 08:22:12.246523 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d2fzh" event={"ID":"007f0b25-94b5-4529-b835-08cdc34e0d6e","Type":"ContainerDied","Data":"96a88818d3c567e4575fc2096f67461bfab333bbf939f7fa14727864172c5f4f"} Sep 30 08:22:13 crc kubenswrapper[4810]: E0930 08:22:13.834077 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Sep 30 08:22:13 crc kubenswrapper[4810]: E0930 08:22:13.834381 4810 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Sep 30 08:22:13 crc kubenswrapper[4810]: E0930 08:22:13.834492 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:38.102.83.41:5001/podified-master-centos10/openstack-barbican-api:watcher_latest,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fdl48,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-k5dlb_openstack(913aa7a2-f310-4b1e-a403-2e5f98b339b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 08:22:13 crc kubenswrapper[4810]: E0930 08:22:13.835624 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-k5dlb" podUID="913aa7a2-f310-4b1e-a403-2e5f98b339b5" Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.963986 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.970916 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.977191 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.981345 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbd8v\" (UniqueName: \"kubernetes.io/projected/cf25d291-3d0d-48c0-9dae-405ef9fd4922-kube-api-access-cbd8v\") pod \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.981420 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55dcn\" (UniqueName: \"kubernetes.io/projected/cb55042a-f0de-40dd-8a25-49a0226f4a8b-kube-api-access-55dcn\") pod \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.981450 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-swift-storage-0\") pod \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.981482 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-scripts\") pod \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.981507 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-config-data\") pod \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.981532 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-sb\") pod \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.981574 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-config\") pod \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.981598 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb55042a-f0de-40dd-8a25-49a0226f4a8b-logs\") pod \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.981709 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-svc\") pod \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.981780 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-nb\") pod \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\" (UID: \"cf25d291-3d0d-48c0-9dae-405ef9fd4922\") " Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.981824 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb55042a-f0de-40dd-8a25-49a0226f4a8b-horizon-secret-key\") pod \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\" (UID: \"cb55042a-f0de-40dd-8a25-49a0226f4a8b\") " Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.982975 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-scripts" (OuterVolumeSpecName: "scripts") pod "cb55042a-f0de-40dd-8a25-49a0226f4a8b" (UID: "cb55042a-f0de-40dd-8a25-49a0226f4a8b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.987513 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb55042a-f0de-40dd-8a25-49a0226f4a8b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "cb55042a-f0de-40dd-8a25-49a0226f4a8b" (UID: "cb55042a-f0de-40dd-8a25-49a0226f4a8b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.988126 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb55042a-f0de-40dd-8a25-49a0226f4a8b-logs" (OuterVolumeSpecName: "logs") pod "cb55042a-f0de-40dd-8a25-49a0226f4a8b" (UID: "cb55042a-f0de-40dd-8a25-49a0226f4a8b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.990639 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-config-data" (OuterVolumeSpecName: "config-data") pod "cb55042a-f0de-40dd-8a25-49a0226f4a8b" (UID: "cb55042a-f0de-40dd-8a25-49a0226f4a8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:13 crc kubenswrapper[4810]: I0930 08:22:13.999178 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf25d291-3d0d-48c0-9dae-405ef9fd4922-kube-api-access-cbd8v" (OuterVolumeSpecName: "kube-api-access-cbd8v") pod "cf25d291-3d0d-48c0-9dae-405ef9fd4922" (UID: "cf25d291-3d0d-48c0-9dae-405ef9fd4922"). InnerVolumeSpecName "kube-api-access-cbd8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.005404 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb55042a-f0de-40dd-8a25-49a0226f4a8b-kube-api-access-55dcn" (OuterVolumeSpecName: "kube-api-access-55dcn") pod "cb55042a-f0de-40dd-8a25-49a0226f4a8b" (UID: "cb55042a-f0de-40dd-8a25-49a0226f4a8b"). InnerVolumeSpecName "kube-api-access-55dcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.049782 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cf25d291-3d0d-48c0-9dae-405ef9fd4922" (UID: "cf25d291-3d0d-48c0-9dae-405ef9fd4922"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.049983 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cf25d291-3d0d-48c0-9dae-405ef9fd4922" (UID: "cf25d291-3d0d-48c0-9dae-405ef9fd4922"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.073900 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cf25d291-3d0d-48c0-9dae-405ef9fd4922" (UID: "cf25d291-3d0d-48c0-9dae-405ef9fd4922"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.079422 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cf25d291-3d0d-48c0-9dae-405ef9fd4922" (UID: "cf25d291-3d0d-48c0-9dae-405ef9fd4922"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.082850 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4bf21836-6124-4093-89ce-5512a9239d73-horizon-secret-key\") pod \"4bf21836-6124-4093-89ce-5512a9239d73\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.082987 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-scripts\") pod \"4bf21836-6124-4093-89ce-5512a9239d73\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.083137 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-config-data\") pod \"4bf21836-6124-4093-89ce-5512a9239d73\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.083217 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ngd6\" (UniqueName: \"kubernetes.io/projected/4bf21836-6124-4093-89ce-5512a9239d73-kube-api-access-8ngd6\") pod \"4bf21836-6124-4093-89ce-5512a9239d73\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.083375 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bf21836-6124-4093-89ce-5512a9239d73-logs\") pod \"4bf21836-6124-4093-89ce-5512a9239d73\" (UID: \"4bf21836-6124-4093-89ce-5512a9239d73\") " Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.083729 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55dcn\" (UniqueName: \"kubernetes.io/projected/cb55042a-f0de-40dd-8a25-49a0226f4a8b-kube-api-access-55dcn\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.083789 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.083849 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.083902 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb55042a-f0de-40dd-8a25-49a0226f4a8b-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.083951 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.084006 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb55042a-f0de-40dd-8a25-49a0226f4a8b-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.084057 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.084126 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.084787 4810 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb55042a-f0de-40dd-8a25-49a0226f4a8b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.084861 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbd8v\" (UniqueName: \"kubernetes.io/projected/cf25d291-3d0d-48c0-9dae-405ef9fd4922-kube-api-access-cbd8v\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.084512 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bf21836-6124-4093-89ce-5512a9239d73-logs" (OuterVolumeSpecName: "logs") pod "4bf21836-6124-4093-89ce-5512a9239d73" (UID: "4bf21836-6124-4093-89ce-5512a9239d73"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.085332 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-scripts" (OuterVolumeSpecName: "scripts") pod "4bf21836-6124-4093-89ce-5512a9239d73" (UID: "4bf21836-6124-4093-89ce-5512a9239d73"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.085609 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-config-data" (OuterVolumeSpecName: "config-data") pod "4bf21836-6124-4093-89ce-5512a9239d73" (UID: "4bf21836-6124-4093-89ce-5512a9239d73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.089676 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bf21836-6124-4093-89ce-5512a9239d73-kube-api-access-8ngd6" (OuterVolumeSpecName: "kube-api-access-8ngd6") pod "4bf21836-6124-4093-89ce-5512a9239d73" (UID: "4bf21836-6124-4093-89ce-5512a9239d73"). InnerVolumeSpecName "kube-api-access-8ngd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.090067 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bf21836-6124-4093-89ce-5512a9239d73-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "4bf21836-6124-4093-89ce-5512a9239d73" (UID: "4bf21836-6124-4093-89ce-5512a9239d73"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.091287 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-config" (OuterVolumeSpecName: "config") pod "cf25d291-3d0d-48c0-9dae-405ef9fd4922" (UID: "cf25d291-3d0d-48c0-9dae-405ef9fd4922"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.186849 4810 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4bf21836-6124-4093-89ce-5512a9239d73-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.186897 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.186907 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bf21836-6124-4093-89ce-5512a9239d73-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.186917 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ngd6\" (UniqueName: \"kubernetes.io/projected/4bf21836-6124-4093-89ce-5512a9239d73-kube-api-access-8ngd6\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.186929 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bf21836-6124-4093-89ce-5512a9239d73-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.186937 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf25d291-3d0d-48c0-9dae-405ef9fd4922-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.262172 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-688b69fc-7dd7f" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.262166 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-688b69fc-7dd7f" event={"ID":"cb55042a-f0de-40dd-8a25-49a0226f4a8b","Type":"ContainerDied","Data":"77bd1b65e7ce45798341795fc0d2d38ad2baa54ee2d5a3c044f3acc694caa2fd"} Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.263890 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" event={"ID":"cf25d291-3d0d-48c0-9dae-405ef9fd4922","Type":"ContainerDied","Data":"1ec309e8eace6dae459c65031b9fe45e3731b2545ed9529bbf89d0515acfa38b"} Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.264048 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.265017 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5648749bd9-8rckr" event={"ID":"4bf21836-6124-4093-89ce-5512a9239d73","Type":"ContainerDied","Data":"e202cb1e140f5f7a0337b544bea017b26339cf6186d267ce83427c97ecdb317c"} Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.265045 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5648749bd9-8rckr" Sep 30 08:22:14 crc kubenswrapper[4810]: E0930 08:22:14.266740 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.41:5001/podified-master-centos10/openstack-barbican-api:watcher_latest\\\"\"" pod="openstack/barbican-db-sync-k5dlb" podUID="913aa7a2-f310-4b1e-a403-2e5f98b339b5" Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.347348 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5648749bd9-8rckr"] Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.354869 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5648749bd9-8rckr"] Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.361521 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b9b87fb55-7jxzz"] Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.387039 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b9b87fb55-7jxzz"] Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.401669 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-688b69fc-7dd7f"] Sep 30 08:22:14 crc kubenswrapper[4810]: I0930 08:22:14.407441 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-688b69fc-7dd7f"] Sep 30 08:22:15 crc kubenswrapper[4810]: E0930 08:22:15.243975 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Sep 30 08:22:15 crc kubenswrapper[4810]: E0930 08:22:15.244015 4810 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.41:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Sep 30 08:22:15 crc kubenswrapper[4810]: E0930 08:22:15.244139 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:38.102.83.41:5001/podified-master-centos10/openstack-cinder-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qggnm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-z7k2l_openstack(e199d2a8-2a58-47c7-8018-13b29c37c2ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 08:22:15 crc kubenswrapper[4810]: E0930 08:22:15.245517 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-z7k2l" podUID="e199d2a8-2a58-47c7-8018-13b29c37c2ad" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.259146 4810 scope.go:117] "RemoveContainer" containerID="cb5133677659ac907189465c3258647ef90e5b32193cd0cf7ce8fb0f5d05ea64" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.286695 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d2fzh" event={"ID":"007f0b25-94b5-4529-b835-08cdc34e0d6e","Type":"ContainerDied","Data":"17407992b4fc4da5bda22c275094a06296d27030eb09a396f535b6414bba7f18"} Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.286753 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17407992b4fc4da5bda22c275094a06296d27030eb09a396f535b6414bba7f18" Sep 30 08:22:15 crc kubenswrapper[4810]: E0930 08:22:15.287326 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.41:5001/podified-master-centos10/openstack-cinder-api:watcher_latest\\\"\"" pod="openstack/cinder-db-sync-z7k2l" podUID="e199d2a8-2a58-47c7-8018-13b29c37c2ad" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.329749 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.347467 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bf21836-6124-4093-89ce-5512a9239d73" path="/var/lib/kubelet/pods/4bf21836-6124-4093-89ce-5512a9239d73/volumes" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.348488 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb55042a-f0de-40dd-8a25-49a0226f4a8b" path="/var/lib/kubelet/pods/cb55042a-f0de-40dd-8a25-49a0226f4a8b/volumes" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.349939 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" path="/var/lib/kubelet/pods/cf25d291-3d0d-48c0-9dae-405ef9fd4922/volumes" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.414906 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-config\") pod \"007f0b25-94b5-4529-b835-08cdc34e0d6e\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.415080 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-combined-ca-bundle\") pod \"007f0b25-94b5-4529-b835-08cdc34e0d6e\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.415182 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmqbx\" (UniqueName: \"kubernetes.io/projected/007f0b25-94b5-4529-b835-08cdc34e0d6e-kube-api-access-qmqbx\") pod \"007f0b25-94b5-4529-b835-08cdc34e0d6e\" (UID: \"007f0b25-94b5-4529-b835-08cdc34e0d6e\") " Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.437881 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/007f0b25-94b5-4529-b835-08cdc34e0d6e-kube-api-access-qmqbx" (OuterVolumeSpecName: "kube-api-access-qmqbx") pod "007f0b25-94b5-4529-b835-08cdc34e0d6e" (UID: "007f0b25-94b5-4529-b835-08cdc34e0d6e"). InnerVolumeSpecName "kube-api-access-qmqbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.446363 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "007f0b25-94b5-4529-b835-08cdc34e0d6e" (UID: "007f0b25-94b5-4529-b835-08cdc34e0d6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.447824 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-config" (OuterVolumeSpecName: "config") pod "007f0b25-94b5-4529-b835-08cdc34e0d6e" (UID: "007f0b25-94b5-4529-b835-08cdc34e0d6e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.517499 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.517550 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/007f0b25-94b5-4529-b835-08cdc34e0d6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.517565 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmqbx\" (UniqueName: \"kubernetes.io/projected/007f0b25-94b5-4529-b835-08cdc34e0d6e-kube-api-access-qmqbx\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.672850 4810 scope.go:117] "RemoveContainer" containerID="11c46bda01fa7f7f62f2e761f2232d7700b63962f2aa93f556d70bd51276e6dd" Sep 30 08:22:15 crc kubenswrapper[4810]: I0930 08:22:15.769582 4810 scope.go:117] "RemoveContainer" containerID="bfcc5073aa9d0282dace4806787c782cc58667fc7ab16add9a1d07edec3d8484" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.062881 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-754cbc7dbb-wbdt6"] Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.099986 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d769bd5f4-zkxj4"] Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.161013 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.171171 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-x57vl"] Sep 30 08:22:16 crc kubenswrapper[4810]: W0930 08:22:16.185462 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacf21e98_01dc_4a20_aa79_bb5549e59f22.slice/crio-7efa99fe45ddcd9d77f1319e461559962e611513928a2cb8ada243ec25b5ebe1 WatchSource:0}: Error finding container 7efa99fe45ddcd9d77f1319e461559962e611513928a2cb8ada243ec25b5ebe1: Status 404 returned error can't find the container with id 7efa99fe45ddcd9d77f1319e461559962e611513928a2cb8ada243ec25b5ebe1 Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.306569 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"277a70ee-91dd-4198-9abd-7e6bb909bec0","Type":"ContainerStarted","Data":"db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb"} Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.313752 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d769bd5f4-zkxj4" event={"ID":"a71102a5-388b-4dcb-bd80-db92ed28f1f5","Type":"ContainerStarted","Data":"c51dfffc199f0e928fdbef45a743a06061156cc86cce277b490e7cb30c7152d5"} Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.319292 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92bcbff2-f0a3-4303-96cc-5f3177731f6a","Type":"ContainerStarted","Data":"fa6601386e8675aa2cb5e215eff6c3f2aa637b95279aa81ba73c2faf6c08b6f4"} Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.325988 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x57vl" event={"ID":"65f3abf6-0015-41fc-9a40-0ed346bb84ec","Type":"ContainerStarted","Data":"4385ebfb02b19188a88d74d576d86e5242883432bd9ddb5d8ed5a83c616e4bb1"} Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.331222 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754cbc7dbb-wbdt6" event={"ID":"c176d78f-1abf-4191-94f8-8ee7be102b18","Type":"ContainerStarted","Data":"c4a402a0d631408a6d207e0bd44625463f0d1368dd2485a3e052ac7d06864225"} Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.334713 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerStarted","Data":"ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45"} Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.334771 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=5.74811379 podStartE2EDuration="33.334661119s" podCreationTimestamp="2025-09-30 08:21:43 +0000 UTC" firstStartedPulling="2025-09-30 08:21:45.565237856 +0000 UTC m=+1129.017437123" lastFinishedPulling="2025-09-30 08:22:13.151785175 +0000 UTC m=+1156.603984452" observedRunningTime="2025-09-30 08:22:16.334614177 +0000 UTC m=+1159.786813444" watchObservedRunningTime="2025-09-30 08:22:16.334661119 +0000 UTC m=+1159.786860386" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.344465 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-98d79fd7c-kz5wm" event={"ID":"dc2cb8f1-4585-43b1-865c-bea9b61ef20c","Type":"ContainerStarted","Data":"02ef2f083fab19cb6ae206c29afed82056cbba7bdb6457c6b4e1f1c10efd387f"} Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.344516 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-98d79fd7c-kz5wm" event={"ID":"dc2cb8f1-4585-43b1-865c-bea9b61ef20c","Type":"ContainerStarted","Data":"96a6199f09eee930d84db591b76f58776c48a86f7139b00970301945fc37beba"} Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.345100 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-98d79fd7c-kz5wm" podUID="dc2cb8f1-4585-43b1-865c-bea9b61ef20c" containerName="horizon-log" containerID="cri-o://96a6199f09eee930d84db591b76f58776c48a86f7139b00970301945fc37beba" gracePeriod=30 Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.345318 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-98d79fd7c-kz5wm" podUID="dc2cb8f1-4585-43b1-865c-bea9b61ef20c" containerName="horizon" containerID="cri-o://02ef2f083fab19cb6ae206c29afed82056cbba7bdb6457c6b4e1f1c10efd387f" gracePeriod=30 Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.356675 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=14.705742691 podStartE2EDuration="33.356657178s" podCreationTimestamp="2025-09-30 08:21:43 +0000 UTC" firstStartedPulling="2025-09-30 08:21:45.536197182 +0000 UTC m=+1128.988396449" lastFinishedPulling="2025-09-30 08:22:04.187111669 +0000 UTC m=+1147.639310936" observedRunningTime="2025-09-30 08:22:16.353778854 +0000 UTC m=+1159.805978121" watchObservedRunningTime="2025-09-30 08:22:16.356657178 +0000 UTC m=+1159.808856445" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.357541 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8xkhj" event={"ID":"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f","Type":"ContainerStarted","Data":"2096808c93bcdbf171afae758a4667a0e204f744f187817e722454267706b358"} Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.359782 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"acf21e98-01dc-4a20-aa79-bb5549e59f22","Type":"ContainerStarted","Data":"7efa99fe45ddcd9d77f1319e461559962e611513928a2cb8ada243ec25b5ebe1"} Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.362163 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d2fzh" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.401930 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-98d79fd7c-kz5wm" podStartSLOduration=3.426165556 podStartE2EDuration="31.401910453s" podCreationTimestamp="2025-09-30 08:21:45 +0000 UTC" firstStartedPulling="2025-09-30 08:21:47.211881642 +0000 UTC m=+1130.664080909" lastFinishedPulling="2025-09-30 08:22:15.187626539 +0000 UTC m=+1158.639825806" observedRunningTime="2025-09-30 08:22:16.371154519 +0000 UTC m=+1159.823353786" watchObservedRunningTime="2025-09-30 08:22:16.401910453 +0000 UTC m=+1159.854109720" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.425847 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-8xkhj" podStartSLOduration=4.262207686 podStartE2EDuration="33.425824218s" podCreationTimestamp="2025-09-30 08:21:43 +0000 UTC" firstStartedPulling="2025-09-30 08:21:46.024049448 +0000 UTC m=+1129.476248715" lastFinishedPulling="2025-09-30 08:22:15.18766594 +0000 UTC m=+1158.639865247" observedRunningTime="2025-09-30 08:22:16.385955269 +0000 UTC m=+1159.838154536" watchObservedRunningTime="2025-09-30 08:22:16.425824218 +0000 UTC m=+1159.878023485" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.567476 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b9b87fb55-7jxzz" podUID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: i/o timeout" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.600810 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675956976c-gd7bn"] Sep 30 08:22:16 crc kubenswrapper[4810]: E0930 08:22:16.601174 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="007f0b25-94b5-4529-b835-08cdc34e0d6e" containerName="neutron-db-sync" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.601186 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="007f0b25-94b5-4529-b835-08cdc34e0d6e" containerName="neutron-db-sync" Sep 30 08:22:16 crc kubenswrapper[4810]: E0930 08:22:16.601213 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerName="dnsmasq-dns" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.601219 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerName="dnsmasq-dns" Sep 30 08:22:16 crc kubenswrapper[4810]: E0930 08:22:16.601228 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerName="init" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.601233 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerName="init" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.601415 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf25d291-3d0d-48c0-9dae-405ef9fd4922" containerName="dnsmasq-dns" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.601431 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="007f0b25-94b5-4529-b835-08cdc34e0d6e" containerName="neutron-db-sync" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.603024 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.645379 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-svc\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.645473 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc28g\" (UniqueName: \"kubernetes.io/projected/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-kube-api-access-nc28g\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.645500 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-nb\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.645596 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-config\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.645615 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-swift-storage-0\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.645630 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-sb\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.646826 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675956976c-gd7bn"] Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.747576 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-config\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.748950 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-swift-storage-0\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.748975 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-sb\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.749044 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-svc\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.749157 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc28g\" (UniqueName: \"kubernetes.io/projected/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-kube-api-access-nc28g\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.749192 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-nb\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.749898 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-config\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.750051 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-nb\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.750890 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-svc\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.757122 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-sb\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.764025 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-swift-storage-0\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.764094 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d889fdd6d-xxkrj"] Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.765649 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.768044 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.775611 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.775630 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-2vm9f" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.775618 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.777677 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d889fdd6d-xxkrj"] Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.783242 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc28g\" (UniqueName: \"kubernetes.io/projected/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-kube-api-access-nc28g\") pod \"dnsmasq-dns-675956976c-gd7bn\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.951673 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-config\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.951746 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-combined-ca-bundle\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.951777 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-ovndb-tls-certs\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.951793 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmndz\" (UniqueName: \"kubernetes.io/projected/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-kube-api-access-zmndz\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.951837 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-httpd-config\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:16 crc kubenswrapper[4810]: I0930 08:22:16.962554 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.061094 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-combined-ca-bundle\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.061457 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-ovndb-tls-certs\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.061476 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmndz\" (UniqueName: \"kubernetes.io/projected/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-kube-api-access-zmndz\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.061547 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-httpd-config\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.061718 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-config\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.064347 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-combined-ca-bundle\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.078124 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-ovndb-tls-certs\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.081461 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-config\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.085718 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmndz\" (UniqueName: \"kubernetes.io/projected/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-kube-api-access-zmndz\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.100129 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-httpd-config\") pod \"neutron-d889fdd6d-xxkrj\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.311719 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.445548 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754cbc7dbb-wbdt6" event={"ID":"c176d78f-1abf-4191-94f8-8ee7be102b18","Type":"ContainerStarted","Data":"65c41b5d112da8a316221d00dc0b08a61accf182470442b32f5aa0e475f2698f"} Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.477801 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"acf21e98-01dc-4a20-aa79-bb5549e59f22","Type":"ContainerStarted","Data":"ce324bf8eb129901d3148020c2f3f5762f1762db9793d57f9d29cd6e47f8cf5f"} Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.506664 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d769bd5f4-zkxj4" event={"ID":"a71102a5-388b-4dcb-bd80-db92ed28f1f5","Type":"ContainerStarted","Data":"c48226a009b4e3562ef85ec916f15fa796d0f1a0fae496619cf58240804d782b"} Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.511744 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x57vl" event={"ID":"65f3abf6-0015-41fc-9a40-0ed346bb84ec","Type":"ContainerStarted","Data":"38f6fd6436640fd7a28907681d9e1a4ae351640a40307e3213e3af643dfa869e"} Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.534555 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4c6qh" event={"ID":"ff6ce927-f393-4895-a7b6-af266138c8c1","Type":"ContainerStarted","Data":"a8483f9453c3170461fd7f633d48cb345a6e19d91c2506f88094b43631ddb18c"} Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.541257 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-x57vl" podStartSLOduration=12.541235468 podStartE2EDuration="12.541235468s" podCreationTimestamp="2025-09-30 08:22:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:17.534492562 +0000 UTC m=+1160.986691829" watchObservedRunningTime="2025-09-30 08:22:17.541235468 +0000 UTC m=+1160.993434755" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.575908 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4c6qh" podStartSLOduration=10.953445021 podStartE2EDuration="1m6.575882715s" podCreationTimestamp="2025-09-30 08:21:11 +0000 UTC" firstStartedPulling="2025-09-30 08:21:18.203648214 +0000 UTC m=+1101.655847481" lastFinishedPulling="2025-09-30 08:22:13.826085898 +0000 UTC m=+1157.278285175" observedRunningTime="2025-09-30 08:22:17.564473184 +0000 UTC m=+1161.016672451" watchObservedRunningTime="2025-09-30 08:22:17.575882715 +0000 UTC m=+1161.028081982" Sep 30 08:22:17 crc kubenswrapper[4810]: I0930 08:22:17.705574 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675956976c-gd7bn"] Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.164035 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d889fdd6d-xxkrj"] Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.412379 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.552216 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754cbc7dbb-wbdt6" event={"ID":"c176d78f-1abf-4191-94f8-8ee7be102b18","Type":"ContainerStarted","Data":"a85450da47bf890fcf9ae454a6f57b37faf86b24f277f113255feb929be07b4c"} Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.579493 4810 generic.go:334] "Generic (PLEG): container finished" podID="06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" containerID="a220086bf118a538d86466e7606c6504bcba26f94ee3793ecd9f4407fc73a027" exitCode=0 Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.579717 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675956976c-gd7bn" event={"ID":"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b","Type":"ContainerDied","Data":"a220086bf118a538d86466e7606c6504bcba26f94ee3793ecd9f4407fc73a027"} Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.579768 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675956976c-gd7bn" event={"ID":"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b","Type":"ContainerStarted","Data":"04e24be23db9db89fa11356285883a2c3b6b25aefd35d153252aa06995827029"} Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.586661 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-754cbc7dbb-wbdt6" podStartSLOduration=26.586640604 podStartE2EDuration="26.586640604s" podCreationTimestamp="2025-09-30 08:21:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:18.572659458 +0000 UTC m=+1162.024858725" watchObservedRunningTime="2025-09-30 08:22:18.586640604 +0000 UTC m=+1162.038839871" Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.590431 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d889fdd6d-xxkrj" event={"ID":"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523","Type":"ContainerStarted","Data":"a6c208086449aaef11826ff90109b226e8838be1ee47c86e591f07b3ca7f743a"} Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.627098 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"acf21e98-01dc-4a20-aa79-bb5549e59f22","Type":"ContainerStarted","Data":"56530e2e9287fd254ca795190034637bb4d0ce4bb1f55e63d8f81b7480eff5d4"} Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.627448 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.652438 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d769bd5f4-zkxj4" event={"ID":"a71102a5-388b-4dcb-bd80-db92ed28f1f5","Type":"ContainerStarted","Data":"e0bfa1aac3fed978ff4beb7a0156ccf9bb2b7e7c5018e04ead554cb6bb9118c2"} Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.657686 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=13.657665657999999 podStartE2EDuration="13.657665658s" podCreationTimestamp="2025-09-30 08:22:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:18.652759885 +0000 UTC m=+1162.104959152" watchObservedRunningTime="2025-09-30 08:22:18.657665658 +0000 UTC m=+1162.109864925" Sep 30 08:22:18 crc kubenswrapper[4810]: I0930 08:22:18.689224 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6d769bd5f4-zkxj4" podStartSLOduration=26.689203694 podStartE2EDuration="26.689203694s" podCreationTimestamp="2025-09-30 08:21:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:18.685213818 +0000 UTC m=+1162.137413085" watchObservedRunningTime="2025-09-30 08:22:18.689203694 +0000 UTC m=+1162.141402961" Sep 30 08:22:19 crc kubenswrapper[4810]: I0930 08:22:19.681333 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d889fdd6d-xxkrj" event={"ID":"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523","Type":"ContainerStarted","Data":"21d2347bf42c6eb44e7ee7a5de628b64b5a98e56ec1865f17862029f6ca75cc7"} Sep 30 08:22:19 crc kubenswrapper[4810]: I0930 08:22:19.681651 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:19 crc kubenswrapper[4810]: I0930 08:22:19.681666 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d889fdd6d-xxkrj" event={"ID":"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523","Type":"ContainerStarted","Data":"d0eeafa4714463a65c8685739ff271f4d79c5316b1d5cc37ad914db6d677ccab"} Sep 30 08:22:19 crc kubenswrapper[4810]: I0930 08:22:19.692105 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675956976c-gd7bn" event={"ID":"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b","Type":"ContainerStarted","Data":"2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778"} Sep 30 08:22:19 crc kubenswrapper[4810]: I0930 08:22:19.700695 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d889fdd6d-xxkrj" podStartSLOduration=3.700670864 podStartE2EDuration="3.700670864s" podCreationTimestamp="2025-09-30 08:22:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:19.694711731 +0000 UTC m=+1163.146911018" watchObservedRunningTime="2025-09-30 08:22:19.700670864 +0000 UTC m=+1163.152870131" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.607627 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.675909 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-675956976c-gd7bn" podStartSLOduration=4.6758918210000004 podStartE2EDuration="4.675891821s" podCreationTimestamp="2025-09-30 08:22:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:19.719128971 +0000 UTC m=+1163.171328238" watchObservedRunningTime="2025-09-30 08:22:20.675891821 +0000 UTC m=+1164.128091088" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.682377 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5f769c46d7-mgh2q"] Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.683984 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.689548 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.689664 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.717982 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f769c46d7-mgh2q"] Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.751627 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.751973 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.859389 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-internal-tls-certs\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.859501 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-public-tls-certs\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.859537 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-combined-ca-bundle\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.859570 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shnn9\" (UniqueName: \"kubernetes.io/projected/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-kube-api-access-shnn9\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.859598 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-config\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.859652 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-httpd-config\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.859667 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-ovndb-tls-certs\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.961376 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-internal-tls-certs\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.961475 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-public-tls-certs\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.961517 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-combined-ca-bundle\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.961543 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shnn9\" (UniqueName: \"kubernetes.io/projected/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-kube-api-access-shnn9\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.961561 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-config\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.961608 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-httpd-config\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.961627 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-ovndb-tls-certs\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.980508 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-combined-ca-bundle\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.980962 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-public-tls-certs\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.981012 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-httpd-config\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.981848 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-ovndb-tls-certs\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.982649 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shnn9\" (UniqueName: \"kubernetes.io/projected/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-kube-api-access-shnn9\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.983196 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-config\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:20 crc kubenswrapper[4810]: I0930 08:22:20.988411 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34fe741b-5b18-4fe8-9f54-1fd3e62d9e24-internal-tls-certs\") pod \"neutron-5f769c46d7-mgh2q\" (UID: \"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24\") " pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:21 crc kubenswrapper[4810]: I0930 08:22:21.053088 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:21 crc kubenswrapper[4810]: I0930 08:22:21.732294 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f769c46d7-mgh2q"] Sep 30 08:22:21 crc kubenswrapper[4810]: W0930 08:22:21.747672 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34fe741b_5b18_4fe8_9f54_1fd3e62d9e24.slice/crio-a91a8bca59026dddd896c90677989c5926292533b0db5c244877376c2c646287 WatchSource:0}: Error finding container a91a8bca59026dddd896c90677989c5926292533b0db5c244877376c2c646287: Status 404 returned error can't find the container with id a91a8bca59026dddd896c90677989c5926292533b0db5c244877376c2c646287 Sep 30 08:22:21 crc kubenswrapper[4810]: I0930 08:22:21.767425 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f769c46d7-mgh2q" event={"ID":"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24","Type":"ContainerStarted","Data":"a91a8bca59026dddd896c90677989c5926292533b0db5c244877376c2c646287"} Sep 30 08:22:21 crc kubenswrapper[4810]: I0930 08:22:21.769740 4810 generic.go:334] "Generic (PLEG): container finished" podID="f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f" containerID="2096808c93bcdbf171afae758a4667a0e204f744f187817e722454267706b358" exitCode=0 Sep 30 08:22:21 crc kubenswrapper[4810]: I0930 08:22:21.770591 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8xkhj" event={"ID":"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f","Type":"ContainerDied","Data":"2096808c93bcdbf171afae758a4667a0e204f744f187817e722454267706b358"} Sep 30 08:22:22 crc kubenswrapper[4810]: I0930 08:22:22.434200 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Sep 30 08:22:22 crc kubenswrapper[4810]: I0930 08:22:22.771900 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:22:22 crc kubenswrapper[4810]: I0930 08:22:22.772860 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:22:22 crc kubenswrapper[4810]: I0930 08:22:22.790162 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f769c46d7-mgh2q" event={"ID":"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24","Type":"ContainerStarted","Data":"c149537b45d8a619e21302cc9013b9d7572085cabec79ab5a4ff79050a57644e"} Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.145531 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.145594 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.226438 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8xkhj" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.241567 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-combined-ca-bundle\") pod \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.241702 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqkqq\" (UniqueName: \"kubernetes.io/projected/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-kube-api-access-bqkqq\") pod \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.241742 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-scripts\") pod \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.241773 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-config-data\") pod \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.241867 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-logs\") pod \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\" (UID: \"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f\") " Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.244623 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-logs" (OuterVolumeSpecName: "logs") pod "f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f" (UID: "f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.253481 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-scripts" (OuterVolumeSpecName: "scripts") pod "f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f" (UID: "f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.255381 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-kube-api-access-bqkqq" (OuterVolumeSpecName: "kube-api-access-bqkqq") pod "f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f" (UID: "f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f"). InnerVolumeSpecName "kube-api-access-bqkqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.281527 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f" (UID: "f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.292350 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-config-data" (OuterVolumeSpecName: "config-data") pod "f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f" (UID: "f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.344448 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.344486 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.344499 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.344511 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.344523 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqkqq\" (UniqueName: \"kubernetes.io/projected/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f-kube-api-access-bqkqq\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.408252 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.456828 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.478324 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:22:23 crc kubenswrapper[4810]: E0930 08:22:23.479095 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45 is running failed: container process not found" containerID="ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Sep 30 08:22:23 crc kubenswrapper[4810]: E0930 08:22:23.479557 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45 is running failed: container process not found" containerID="ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Sep 30 08:22:23 crc kubenswrapper[4810]: E0930 08:22:23.481174 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45 is running failed: container process not found" containerID="ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Sep 30 08:22:23 crc kubenswrapper[4810]: E0930 08:22:23.481224 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45 is running failed: container process not found" probeType="Startup" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.801354 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f769c46d7-mgh2q" event={"ID":"34fe741b-5b18-4fe8-9f54-1fd3e62d9e24","Type":"ContainerStarted","Data":"8b31ace1da483578797027181be45f42ca1d1e8fd1af4ef2cbe140f8f85d1a0e"} Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.801593 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.804047 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8xkhj" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.805013 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8xkhj" event={"ID":"f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f","Type":"ContainerDied","Data":"5b8a148dc3a7f0c593194596567d8de523c4b7dbb0f85955593ad2971bd011a7"} Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.805064 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b8a148dc3a7f0c593194596567d8de523c4b7dbb0f85955593ad2971bd011a7" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.806871 4810 generic.go:334] "Generic (PLEG): container finished" podID="65f3abf6-0015-41fc-9a40-0ed346bb84ec" containerID="38f6fd6436640fd7a28907681d9e1a4ae351640a40307e3213e3af643dfa869e" exitCode=0 Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.806949 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x57vl" event={"ID":"65f3abf6-0015-41fc-9a40-0ed346bb84ec","Type":"ContainerDied","Data":"38f6fd6436640fd7a28907681d9e1a4ae351640a40307e3213e3af643dfa869e"} Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.808709 4810 generic.go:334] "Generic (PLEG): container finished" podID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerID="ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45" exitCode=1 Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.808813 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerDied","Data":"ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45"} Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.809713 4810 scope.go:117] "RemoveContainer" containerID="ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.855925 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5f769c46d7-mgh2q" podStartSLOduration=3.855899092 podStartE2EDuration="3.855899092s" podCreationTimestamp="2025-09-30 08:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:23.82689264 +0000 UTC m=+1167.279091917" watchObservedRunningTime="2025-09-30 08:22:23.855899092 +0000 UTC m=+1167.308098359" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.861503 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.920575 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-54b65d888b-2hkpr"] Sep 30 08:22:23 crc kubenswrapper[4810]: E0930 08:22:23.925127 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f" containerName="placement-db-sync" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.925147 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f" containerName="placement-db-sync" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.925388 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f" containerName="placement-db-sync" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.926382 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.929713 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.930000 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-szzm6" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.930146 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.930258 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.938141 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.961609 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-internal-tls-certs\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.961658 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q8c8\" (UniqueName: \"kubernetes.io/projected/35128b6f-4e9c-414b-847f-773c1418f5e2-kube-api-access-6q8c8\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.961710 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-scripts\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.961743 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-public-tls-certs\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.961812 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-combined-ca-bundle\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.961829 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35128b6f-4e9c-414b-847f-773c1418f5e2-logs\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.961850 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-config-data\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:23 crc kubenswrapper[4810]: I0930 08:22:23.987923 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-54b65d888b-2hkpr"] Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.022708 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.066482 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-scripts\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.066551 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-public-tls-certs\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.066597 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-combined-ca-bundle\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.066614 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35128b6f-4e9c-414b-847f-773c1418f5e2-logs\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.066636 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-config-data\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.066723 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-internal-tls-certs\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.066746 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q8c8\" (UniqueName: \"kubernetes.io/projected/35128b6f-4e9c-414b-847f-773c1418f5e2-kube-api-access-6q8c8\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.067781 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35128b6f-4e9c-414b-847f-773c1418f5e2-logs\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.070882 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-scripts\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.075127 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-config-data\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.079688 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-internal-tls-certs\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.079826 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-public-tls-certs\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.088815 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35128b6f-4e9c-414b-847f-773c1418f5e2-combined-ca-bundle\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.090313 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q8c8\" (UniqueName: \"kubernetes.io/projected/35128b6f-4e9c-414b-847f-773c1418f5e2-kube-api-access-6q8c8\") pod \"placement-54b65d888b-2hkpr\" (UID: \"35128b6f-4e9c-414b-847f-773c1418f5e2\") " pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.253891 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.759355 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-54b65d888b-2hkpr"] Sep 30 08:22:24 crc kubenswrapper[4810]: W0930 08:22:24.762183 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35128b6f_4e9c_414b_847f_773c1418f5e2.slice/crio-2c8d4b764a12e9facb7ae3c66655e0d9adcc642131042dc924a4d493fdfd5a91 WatchSource:0}: Error finding container 2c8d4b764a12e9facb7ae3c66655e0d9adcc642131042dc924a4d493fdfd5a91: Status 404 returned error can't find the container with id 2c8d4b764a12e9facb7ae3c66655e0d9adcc642131042dc924a4d493fdfd5a91 Sep 30 08:22:24 crc kubenswrapper[4810]: I0930 08:22:24.834923 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54b65d888b-2hkpr" event={"ID":"35128b6f-4e9c-414b-847f-773c1418f5e2","Type":"ContainerStarted","Data":"2c8d4b764a12e9facb7ae3c66655e0d9adcc642131042dc924a4d493fdfd5a91"} Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.283663 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.394070 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-credential-keys\") pod \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.394280 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-config-data\") pod \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.394397 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-fernet-keys\") pod \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.394478 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t5pn\" (UniqueName: \"kubernetes.io/projected/65f3abf6-0015-41fc-9a40-0ed346bb84ec-kube-api-access-4t5pn\") pod \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.394529 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-scripts\") pod \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.394557 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-combined-ca-bundle\") pod \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\" (UID: \"65f3abf6-0015-41fc-9a40-0ed346bb84ec\") " Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.400929 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65f3abf6-0015-41fc-9a40-0ed346bb84ec-kube-api-access-4t5pn" (OuterVolumeSpecName: "kube-api-access-4t5pn") pod "65f3abf6-0015-41fc-9a40-0ed346bb84ec" (UID: "65f3abf6-0015-41fc-9a40-0ed346bb84ec"). InnerVolumeSpecName "kube-api-access-4t5pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.403396 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "65f3abf6-0015-41fc-9a40-0ed346bb84ec" (UID: "65f3abf6-0015-41fc-9a40-0ed346bb84ec"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.404945 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "65f3abf6-0015-41fc-9a40-0ed346bb84ec" (UID: "65f3abf6-0015-41fc-9a40-0ed346bb84ec"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.405385 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-scripts" (OuterVolumeSpecName: "scripts") pod "65f3abf6-0015-41fc-9a40-0ed346bb84ec" (UID: "65f3abf6-0015-41fc-9a40-0ed346bb84ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.426623 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65f3abf6-0015-41fc-9a40-0ed346bb84ec" (UID: "65f3abf6-0015-41fc-9a40-0ed346bb84ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.453907 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-config-data" (OuterVolumeSpecName: "config-data") pod "65f3abf6-0015-41fc-9a40-0ed346bb84ec" (UID: "65f3abf6-0015-41fc-9a40-0ed346bb84ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.500119 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.500159 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.500175 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t5pn\" (UniqueName: \"kubernetes.io/projected/65f3abf6-0015-41fc-9a40-0ed346bb84ec-kube-api-access-4t5pn\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.500190 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.500205 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.500218 4810 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/65f3abf6-0015-41fc-9a40-0ed346bb84ec-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.606723 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.612918 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.842840 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x57vl" event={"ID":"65f3abf6-0015-41fc-9a40-0ed346bb84ec","Type":"ContainerDied","Data":"4385ebfb02b19188a88d74d576d86e5242883432bd9ddb5d8ed5a83c616e4bb1"} Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.842893 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4385ebfb02b19188a88d74d576d86e5242883432bd9ddb5d8ed5a83c616e4bb1" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.842951 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-applier-0" podUID="277a70ee-91dd-4198-9abd-7e6bb909bec0" containerName="watcher-applier" containerID="cri-o://db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb" gracePeriod=30 Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.843046 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x57vl" Sep 30 08:22:25 crc kubenswrapper[4810]: I0930 08:22:25.851459 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.028801 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f95cc94f6-tz5ks"] Sep 30 08:22:26 crc kubenswrapper[4810]: E0930 08:22:26.029289 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f3abf6-0015-41fc-9a40-0ed346bb84ec" containerName="keystone-bootstrap" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.029310 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f3abf6-0015-41fc-9a40-0ed346bb84ec" containerName="keystone-bootstrap" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.029561 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f3abf6-0015-41fc-9a40-0ed346bb84ec" containerName="keystone-bootstrap" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.031904 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.039220 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.049436 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f95cc94f6-tz5ks"] Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.049962 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.050621 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.050664 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.050754 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.050793 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-hw6qv" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.114799 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-scripts\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.114862 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-public-tls-certs\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.114908 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-internal-tls-certs\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.114927 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-fernet-keys\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.114974 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shdq5\" (UniqueName: \"kubernetes.io/projected/2d05e909-d2bc-49d7-af7b-0736de4c3f95-kube-api-access-shdq5\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.115051 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-credential-keys\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.115088 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-config-data\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.115110 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-combined-ca-bundle\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.216663 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-config-data\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.216720 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-combined-ca-bundle\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.216994 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-scripts\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.217451 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-public-tls-certs\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.217520 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-internal-tls-certs\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.217539 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-fernet-keys\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.217588 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shdq5\" (UniqueName: \"kubernetes.io/projected/2d05e909-d2bc-49d7-af7b-0736de4c3f95-kube-api-access-shdq5\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.217671 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-credential-keys\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.222949 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-internal-tls-certs\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.226950 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-config-data\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.226985 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-fernet-keys\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.227138 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-scripts\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.227450 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-public-tls-certs\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.227500 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-credential-keys\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.230979 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d05e909-d2bc-49d7-af7b-0736de4c3f95-combined-ca-bundle\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.235116 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shdq5\" (UniqueName: \"kubernetes.io/projected/2d05e909-d2bc-49d7-af7b-0736de4c3f95-kube-api-access-shdq5\") pod \"keystone-f95cc94f6-tz5ks\" (UID: \"2d05e909-d2bc-49d7-af7b-0736de4c3f95\") " pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.329130 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.363828 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.881603 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f95cc94f6-tz5ks"] Sep 30 08:22:26 crc kubenswrapper[4810]: W0930 08:22:26.885028 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d05e909_d2bc_49d7_af7b_0736de4c3f95.slice/crio-2a1b6cef2da330d81f1cf655fc55036c843ff19758b82985686ed68196bbcd23 WatchSource:0}: Error finding container 2a1b6cef2da330d81f1cf655fc55036c843ff19758b82985686ed68196bbcd23: Status 404 returned error can't find the container with id 2a1b6cef2da330d81f1cf655fc55036c843ff19758b82985686ed68196bbcd23 Sep 30 08:22:26 crc kubenswrapper[4810]: I0930 08:22:26.963428 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:27 crc kubenswrapper[4810]: I0930 08:22:27.038898 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-548f9979d5-zkcjj"] Sep 30 08:22:27 crc kubenswrapper[4810]: I0930 08:22:27.039236 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" podUID="0f887e1a-55da-42cd-a8f1-709f797b283e" containerName="dnsmasq-dns" containerID="cri-o://6a26e395aab71c0f229066ce5c97280e9ba8cb043f774c09c05d83df7239c0b4" gracePeriod=10 Sep 30 08:22:27 crc kubenswrapper[4810]: E0930 08:22:27.492961 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f887e1a_55da_42cd_a8f1_709f797b283e.slice/crio-6a26e395aab71c0f229066ce5c97280e9ba8cb043f774c09c05d83df7239c0b4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f887e1a_55da_42cd_a8f1_709f797b283e.slice/crio-conmon-6a26e395aab71c0f229066ce5c97280e9ba8cb043f774c09c05d83df7239c0b4.scope\": RecentStats: unable to find data in memory cache]" Sep 30 08:22:27 crc kubenswrapper[4810]: I0930 08:22:27.863937 4810 generic.go:334] "Generic (PLEG): container finished" podID="0f887e1a-55da-42cd-a8f1-709f797b283e" containerID="6a26e395aab71c0f229066ce5c97280e9ba8cb043f774c09c05d83df7239c0b4" exitCode=0 Sep 30 08:22:27 crc kubenswrapper[4810]: I0930 08:22:27.864017 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" event={"ID":"0f887e1a-55da-42cd-a8f1-709f797b283e","Type":"ContainerDied","Data":"6a26e395aab71c0f229066ce5c97280e9ba8cb043f774c09c05d83df7239c0b4"} Sep 30 08:22:27 crc kubenswrapper[4810]: I0930 08:22:27.865173 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f95cc94f6-tz5ks" event={"ID":"2d05e909-d2bc-49d7-af7b-0736de4c3f95","Type":"ContainerStarted","Data":"2a1b6cef2da330d81f1cf655fc55036c843ff19758b82985686ed68196bbcd23"} Sep 30 08:22:28 crc kubenswrapper[4810]: E0930 08:22:28.409204 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Sep 30 08:22:28 crc kubenswrapper[4810]: E0930 08:22:28.410470 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Sep 30 08:22:28 crc kubenswrapper[4810]: E0930 08:22:28.411792 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Sep 30 08:22:28 crc kubenswrapper[4810]: E0930 08:22:28.411837 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="277a70ee-91dd-4198-9abd-7e6bb909bec0" containerName="watcher-applier" Sep 30 08:22:28 crc kubenswrapper[4810]: I0930 08:22:28.894765 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54b65d888b-2hkpr" event={"ID":"35128b6f-4e9c-414b-847f-773c1418f5e2","Type":"ContainerStarted","Data":"f402b5e98b52d17a6ad6804f199dd1c3196a3f99cec241803695a6b3915618b4"} Sep 30 08:22:28 crc kubenswrapper[4810]: I0930 08:22:28.992364 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:22:28 crc kubenswrapper[4810]: I0930 08:22:28.992635 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerName="watcher-api-log" containerID="cri-o://ce324bf8eb129901d3148020c2f3f5762f1762db9793d57f9d29cd6e47f8cf5f" gracePeriod=30 Sep 30 08:22:28 crc kubenswrapper[4810]: I0930 08:22:28.992687 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerName="watcher-api" containerID="cri-o://56530e2e9287fd254ca795190034637bb4d0ce4bb1f55e63d8f81b7480eff5d4" gracePeriod=30 Sep 30 08:22:29 crc kubenswrapper[4810]: I0930 08:22:29.178322 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" podUID="0f887e1a-55da-42cd-a8f1-709f797b283e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.161:5353: connect: connection refused" Sep 30 08:22:29 crc kubenswrapper[4810]: I0930 08:22:29.905872 4810 generic.go:334] "Generic (PLEG): container finished" podID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerID="ce324bf8eb129901d3148020c2f3f5762f1762db9793d57f9d29cd6e47f8cf5f" exitCode=143 Sep 30 08:22:29 crc kubenswrapper[4810]: I0930 08:22:29.906050 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"acf21e98-01dc-4a20-aa79-bb5549e59f22","Type":"ContainerDied","Data":"ce324bf8eb129901d3148020c2f3f5762f1762db9793d57f9d29cd6e47f8cf5f"} Sep 30 08:22:29 crc kubenswrapper[4810]: I0930 08:22:29.914819 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f95cc94f6-tz5ks" event={"ID":"2d05e909-d2bc-49d7-af7b-0736de4c3f95","Type":"ContainerStarted","Data":"a7e9a6bff0e78a29f8fea7f29b1e60cc8243f0223c62e9fe7820e149566ecd1f"} Sep 30 08:22:29 crc kubenswrapper[4810]: I0930 08:22:29.914983 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:22:29 crc kubenswrapper[4810]: I0930 08:22:29.922745 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerStarted","Data":"1f7b36decfbc46430a048f59252a01883a015411e53ee22a10670b789efa6fb8"} Sep 30 08:22:29 crc kubenswrapper[4810]: I0930 08:22:29.930975 4810 generic.go:334] "Generic (PLEG): container finished" podID="277a70ee-91dd-4198-9abd-7e6bb909bec0" containerID="db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb" exitCode=0 Sep 30 08:22:29 crc kubenswrapper[4810]: I0930 08:22:29.931027 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"277a70ee-91dd-4198-9abd-7e6bb909bec0","Type":"ContainerDied","Data":"db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb"} Sep 30 08:22:29 crc kubenswrapper[4810]: I0930 08:22:29.933830 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-f95cc94f6-tz5ks" podStartSLOduration=4.933809877 podStartE2EDuration="4.933809877s" podCreationTimestamp="2025-09-30 08:22:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:29.932658123 +0000 UTC m=+1173.384857390" watchObservedRunningTime="2025-09-30 08:22:29.933809877 +0000 UTC m=+1173.386009144" Sep 30 08:22:30 crc kubenswrapper[4810]: I0930 08:22:30.607258 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.165:9322/\": dial tcp 10.217.0.165:9322: connect: connection refused" Sep 30 08:22:30 crc kubenswrapper[4810]: I0930 08:22:30.607609 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.165:9322/\": dial tcp 10.217.0.165:9322: connect: connection refused" Sep 30 08:22:30 crc kubenswrapper[4810]: I0930 08:22:30.942948 4810 generic.go:334] "Generic (PLEG): container finished" podID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerID="56530e2e9287fd254ca795190034637bb4d0ce4bb1f55e63d8f81b7480eff5d4" exitCode=0 Sep 30 08:22:30 crc kubenswrapper[4810]: I0930 08:22:30.943069 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"acf21e98-01dc-4a20-aa79-bb5549e59f22","Type":"ContainerDied","Data":"56530e2e9287fd254ca795190034637bb4d0ce4bb1f55e63d8f81b7480eff5d4"} Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.431825 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.562154 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-config\") pod \"0f887e1a-55da-42cd-a8f1-709f797b283e\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.562501 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-sb\") pod \"0f887e1a-55da-42cd-a8f1-709f797b283e\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.562543 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2h6n\" (UniqueName: \"kubernetes.io/projected/0f887e1a-55da-42cd-a8f1-709f797b283e-kube-api-access-r2h6n\") pod \"0f887e1a-55da-42cd-a8f1-709f797b283e\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.562584 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-svc\") pod \"0f887e1a-55da-42cd-a8f1-709f797b283e\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.562640 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-swift-storage-0\") pod \"0f887e1a-55da-42cd-a8f1-709f797b283e\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.562741 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-nb\") pod \"0f887e1a-55da-42cd-a8f1-709f797b283e\" (UID: \"0f887e1a-55da-42cd-a8f1-709f797b283e\") " Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.569634 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f887e1a-55da-42cd-a8f1-709f797b283e-kube-api-access-r2h6n" (OuterVolumeSpecName: "kube-api-access-r2h6n") pod "0f887e1a-55da-42cd-a8f1-709f797b283e" (UID: "0f887e1a-55da-42cd-a8f1-709f797b283e"). InnerVolumeSpecName "kube-api-access-r2h6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.629149 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-config" (OuterVolumeSpecName: "config") pod "0f887e1a-55da-42cd-a8f1-709f797b283e" (UID: "0f887e1a-55da-42cd-a8f1-709f797b283e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.632842 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0f887e1a-55da-42cd-a8f1-709f797b283e" (UID: "0f887e1a-55da-42cd-a8f1-709f797b283e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.636676 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0f887e1a-55da-42cd-a8f1-709f797b283e" (UID: "0f887e1a-55da-42cd-a8f1-709f797b283e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.640143 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0f887e1a-55da-42cd-a8f1-709f797b283e" (UID: "0f887e1a-55da-42cd-a8f1-709f797b283e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.642575 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0f887e1a-55da-42cd-a8f1-709f797b283e" (UID: "0f887e1a-55da-42cd-a8f1-709f797b283e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.665029 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.665076 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.665120 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.665138 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2h6n\" (UniqueName: \"kubernetes.io/projected/0f887e1a-55da-42cd-a8f1-709f797b283e-kube-api-access-r2h6n\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.665154 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.665166 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f887e1a-55da-42cd-a8f1-709f797b283e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.961167 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" event={"ID":"0f887e1a-55da-42cd-a8f1-709f797b283e","Type":"ContainerDied","Data":"3f84a89dda6951f78d490d5b349989c1beba322fae674b13b9f7661976126240"} Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.961314 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548f9979d5-zkcjj" Sep 30 08:22:31 crc kubenswrapper[4810]: I0930 08:22:31.962063 4810 scope.go:117] "RemoveContainer" containerID="6a26e395aab71c0f229066ce5c97280e9ba8cb043f774c09c05d83df7239c0b4" Sep 30 08:22:32 crc kubenswrapper[4810]: I0930 08:22:32.002724 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-548f9979d5-zkcjj"] Sep 30 08:22:32 crc kubenswrapper[4810]: I0930 08:22:32.014485 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-548f9979d5-zkcjj"] Sep 30 08:22:32 crc kubenswrapper[4810]: I0930 08:22:32.970326 4810 generic.go:334] "Generic (PLEG): container finished" podID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerID="1f7b36decfbc46430a048f59252a01883a015411e53ee22a10670b789efa6fb8" exitCode=1 Sep 30 08:22:32 crc kubenswrapper[4810]: I0930 08:22:32.970410 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerDied","Data":"1f7b36decfbc46430a048f59252a01883a015411e53ee22a10670b789efa6fb8"} Sep 30 08:22:32 crc kubenswrapper[4810]: I0930 08:22:32.971074 4810 scope.go:117] "RemoveContainer" containerID="1f7b36decfbc46430a048f59252a01883a015411e53ee22a10670b789efa6fb8" Sep 30 08:22:32 crc kubenswrapper[4810]: E0930 08:22:32.971393 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:22:33 crc kubenswrapper[4810]: I0930 08:22:33.318381 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f887e1a-55da-42cd-a8f1-709f797b283e" path="/var/lib/kubelet/pods/0f887e1a-55da-42cd-a8f1-709f797b283e/volumes" Sep 30 08:22:33 crc kubenswrapper[4810]: E0930 08:22:33.409042 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb is running failed: container process not found" containerID="db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Sep 30 08:22:33 crc kubenswrapper[4810]: E0930 08:22:33.409658 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb is running failed: container process not found" containerID="db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Sep 30 08:22:33 crc kubenswrapper[4810]: E0930 08:22:33.410374 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb is running failed: container process not found" containerID="db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Sep 30 08:22:33 crc kubenswrapper[4810]: E0930 08:22:33.410414 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb is running failed: container process not found" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="277a70ee-91dd-4198-9abd-7e6bb909bec0" containerName="watcher-applier" Sep 30 08:22:33 crc kubenswrapper[4810]: I0930 08:22:33.478489 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:22:33 crc kubenswrapper[4810]: I0930 08:22:33.478558 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:22:33 crc kubenswrapper[4810]: I0930 08:22:33.980415 4810 scope.go:117] "RemoveContainer" containerID="1f7b36decfbc46430a048f59252a01883a015411e53ee22a10670b789efa6fb8" Sep 30 08:22:33 crc kubenswrapper[4810]: E0930 08:22:33.980741 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:22:34 crc kubenswrapper[4810]: I0930 08:22:34.567397 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:22:34 crc kubenswrapper[4810]: I0930 08:22:34.922662 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.038414 4810 scope.go:117] "RemoveContainer" containerID="e52ada73ebee88d08bce03eb187c0069f0d9fd0c6a24f068dc26fd3191ace16b" Sep 30 08:22:35 crc kubenswrapper[4810]: E0930 08:22:35.051721 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/sg-core:latest" Sep 30 08:22:35 crc kubenswrapper[4810]: E0930 08:22:35.051889 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:sg-core,Image:quay.io/openstack-k8s-operators/sg-core:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sg-core-conf-yaml,ReadOnly:false,MountPath:/etc/sg-core.conf.yaml,SubPath:sg-core.conf.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qmx6l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(92bcbff2-f0a3-4303-96cc-5f3177731f6a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.112038 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.119385 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.197452 4810 scope.go:117] "RemoveContainer" containerID="ef70cc9bec40c302a12aab6526801dbc261074afa26ebb944e7755a431355d45" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.266324 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-config-data\") pod \"277a70ee-91dd-4198-9abd-7e6bb909bec0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.266408 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acf21e98-01dc-4a20-aa79-bb5549e59f22-logs\") pod \"acf21e98-01dc-4a20-aa79-bb5549e59f22\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.266453 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-combined-ca-bundle\") pod \"acf21e98-01dc-4a20-aa79-bb5549e59f22\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.266476 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snxfp\" (UniqueName: \"kubernetes.io/projected/acf21e98-01dc-4a20-aa79-bb5549e59f22-kube-api-access-snxfp\") pod \"acf21e98-01dc-4a20-aa79-bb5549e59f22\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.266530 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-combined-ca-bundle\") pod \"277a70ee-91dd-4198-9abd-7e6bb909bec0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.266582 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59pdh\" (UniqueName: \"kubernetes.io/projected/277a70ee-91dd-4198-9abd-7e6bb909bec0-kube-api-access-59pdh\") pod \"277a70ee-91dd-4198-9abd-7e6bb909bec0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.266623 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-custom-prometheus-ca\") pod \"acf21e98-01dc-4a20-aa79-bb5549e59f22\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.266692 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-config-data\") pod \"acf21e98-01dc-4a20-aa79-bb5549e59f22\" (UID: \"acf21e98-01dc-4a20-aa79-bb5549e59f22\") " Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.266718 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/277a70ee-91dd-4198-9abd-7e6bb909bec0-logs\") pod \"277a70ee-91dd-4198-9abd-7e6bb909bec0\" (UID: \"277a70ee-91dd-4198-9abd-7e6bb909bec0\") " Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.268008 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/277a70ee-91dd-4198-9abd-7e6bb909bec0-logs" (OuterVolumeSpecName: "logs") pod "277a70ee-91dd-4198-9abd-7e6bb909bec0" (UID: "277a70ee-91dd-4198-9abd-7e6bb909bec0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.268762 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acf21e98-01dc-4a20-aa79-bb5549e59f22-logs" (OuterVolumeSpecName: "logs") pod "acf21e98-01dc-4a20-aa79-bb5549e59f22" (UID: "acf21e98-01dc-4a20-aa79-bb5549e59f22"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.274243 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/277a70ee-91dd-4198-9abd-7e6bb909bec0-kube-api-access-59pdh" (OuterVolumeSpecName: "kube-api-access-59pdh") pod "277a70ee-91dd-4198-9abd-7e6bb909bec0" (UID: "277a70ee-91dd-4198-9abd-7e6bb909bec0"). InnerVolumeSpecName "kube-api-access-59pdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.281177 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acf21e98-01dc-4a20-aa79-bb5549e59f22-kube-api-access-snxfp" (OuterVolumeSpecName: "kube-api-access-snxfp") pod "acf21e98-01dc-4a20-aa79-bb5549e59f22" (UID: "acf21e98-01dc-4a20-aa79-bb5549e59f22"). InnerVolumeSpecName "kube-api-access-snxfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.303881 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "acf21e98-01dc-4a20-aa79-bb5549e59f22" (UID: "acf21e98-01dc-4a20-aa79-bb5549e59f22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.304981 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "acf21e98-01dc-4a20-aa79-bb5549e59f22" (UID: "acf21e98-01dc-4a20-aa79-bb5549e59f22"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.369588 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/277a70ee-91dd-4198-9abd-7e6bb909bec0-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.369622 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acf21e98-01dc-4a20-aa79-bb5549e59f22-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.369631 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.369639 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snxfp\" (UniqueName: \"kubernetes.io/projected/acf21e98-01dc-4a20-aa79-bb5549e59f22-kube-api-access-snxfp\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.369651 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59pdh\" (UniqueName: \"kubernetes.io/projected/277a70ee-91dd-4198-9abd-7e6bb909bec0-kube-api-access-59pdh\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.369659 4810 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.381701 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "277a70ee-91dd-4198-9abd-7e6bb909bec0" (UID: "277a70ee-91dd-4198-9abd-7e6bb909bec0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.398348 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-config-data" (OuterVolumeSpecName: "config-data") pod "277a70ee-91dd-4198-9abd-7e6bb909bec0" (UID: "277a70ee-91dd-4198-9abd-7e6bb909bec0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.423795 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-config-data" (OuterVolumeSpecName: "config-data") pod "acf21e98-01dc-4a20-aa79-bb5549e59f22" (UID: "acf21e98-01dc-4a20-aa79-bb5549e59f22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.471491 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.471526 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf21e98-01dc-4a20-aa79-bb5549e59f22-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:35 crc kubenswrapper[4810]: I0930 08:22:35.471535 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277a70ee-91dd-4198-9abd-7e6bb909bec0-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.001479 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"acf21e98-01dc-4a20-aa79-bb5549e59f22","Type":"ContainerDied","Data":"7efa99fe45ddcd9d77f1319e461559962e611513928a2cb8ada243ec25b5ebe1"} Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.001513 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.001534 4810 scope.go:117] "RemoveContainer" containerID="56530e2e9287fd254ca795190034637bb4d0ce4bb1f55e63d8f81b7480eff5d4" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.003193 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k5dlb" event={"ID":"913aa7a2-f310-4b1e-a403-2e5f98b339b5","Type":"ContainerStarted","Data":"f0abbb76c1572c27e8547117cdc22146c0ce28960bd84e36efdb24d3e672034c"} Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.019787 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"277a70ee-91dd-4198-9abd-7e6bb909bec0","Type":"ContainerDied","Data":"4f4094cc21166c6514efd67967a1643035a9b7d43b18dfee4882fcd89e6c90b6"} Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.019882 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.021288 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-k5dlb" podStartSLOduration=3.293181828 podStartE2EDuration="53.02127886s" podCreationTimestamp="2025-09-30 08:21:43 +0000 UTC" firstStartedPulling="2025-09-30 08:21:45.511781833 +0000 UTC m=+1128.963981100" lastFinishedPulling="2025-09-30 08:22:35.239878875 +0000 UTC m=+1178.692078132" observedRunningTime="2025-09-30 08:22:36.017050967 +0000 UTC m=+1179.469250234" watchObservedRunningTime="2025-09-30 08:22:36.02127886 +0000 UTC m=+1179.473478127" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.046610 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54b65d888b-2hkpr" event={"ID":"35128b6f-4e9c-414b-847f-773c1418f5e2","Type":"ContainerStarted","Data":"e1b9ad6efdf425f661a1be2a4c5a90d2535bd0070f5353ed568b880537283b1e"} Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.048419 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.048716 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.071355 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.084343 4810 scope.go:117] "RemoveContainer" containerID="ce324bf8eb129901d3148020c2f3f5762f1762db9793d57f9d29cd6e47f8cf5f" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.102035 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.117006 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:22:36 crc kubenswrapper[4810]: E0930 08:22:36.117395 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerName="watcher-api" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.117412 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerName="watcher-api" Sep 30 08:22:36 crc kubenswrapper[4810]: E0930 08:22:36.117424 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f887e1a-55da-42cd-a8f1-709f797b283e" containerName="init" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.117430 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f887e1a-55da-42cd-a8f1-709f797b283e" containerName="init" Sep 30 08:22:36 crc kubenswrapper[4810]: E0930 08:22:36.117443 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277a70ee-91dd-4198-9abd-7e6bb909bec0" containerName="watcher-applier" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.117449 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="277a70ee-91dd-4198-9abd-7e6bb909bec0" containerName="watcher-applier" Sep 30 08:22:36 crc kubenswrapper[4810]: E0930 08:22:36.117466 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerName="watcher-api-log" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.117472 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerName="watcher-api-log" Sep 30 08:22:36 crc kubenswrapper[4810]: E0930 08:22:36.117496 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f887e1a-55da-42cd-a8f1-709f797b283e" containerName="dnsmasq-dns" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.117503 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f887e1a-55da-42cd-a8f1-709f797b283e" containerName="dnsmasq-dns" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.117672 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f887e1a-55da-42cd-a8f1-709f797b283e" containerName="dnsmasq-dns" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.117683 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerName="watcher-api-log" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.117699 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="277a70ee-91dd-4198-9abd-7e6bb909bec0" containerName="watcher-applier" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.117707 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="acf21e98-01dc-4a20-aa79-bb5549e59f22" containerName="watcher-api" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.118746 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.121753 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.122545 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.122733 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.138985 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.139549 4810 scope.go:117] "RemoveContainer" containerID="db72bc356adba44d1dcbf0d8e12f2b348bdbc84f626e0bd6755695ccaaf6e2fb" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.146396 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-54b65d888b-2hkpr" podStartSLOduration=13.146375605 podStartE2EDuration="13.146375605s" podCreationTimestamp="2025-09-30 08:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:36.078025129 +0000 UTC m=+1179.530224396" watchObservedRunningTime="2025-09-30 08:22:36.146375605 +0000 UTC m=+1179.598574862" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.170331 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.176736 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-applier-0"] Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.201105 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.203137 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.206559 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.240508 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.296225 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04899be-cfe2-4870-ad3c-735148e7bcdd-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.296535 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32ede37e-d419-4c76-932e-af1de102a9ad-logs\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.296597 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-925bd\" (UniqueName: \"kubernetes.io/projected/d04899be-cfe2-4870-ad3c-735148e7bcdd-kube-api-access-925bd\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.296622 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlff9\" (UniqueName: \"kubernetes.io/projected/32ede37e-d419-4c76-932e-af1de102a9ad-kube-api-access-hlff9\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.296646 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-public-tls-certs\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.296686 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.296703 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04899be-cfe2-4870-ad3c-735148e7bcdd-config-data\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.296729 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.296774 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.296793 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-config-data\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.296817 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d04899be-cfe2-4870-ad3c-735148e7bcdd-logs\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.398737 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d04899be-cfe2-4870-ad3c-735148e7bcdd-logs\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.398862 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04899be-cfe2-4870-ad3c-735148e7bcdd-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.398897 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32ede37e-d419-4c76-932e-af1de102a9ad-logs\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.398982 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-925bd\" (UniqueName: \"kubernetes.io/projected/d04899be-cfe2-4870-ad3c-735148e7bcdd-kube-api-access-925bd\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.399016 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlff9\" (UniqueName: \"kubernetes.io/projected/32ede37e-d419-4c76-932e-af1de102a9ad-kube-api-access-hlff9\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.399068 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-public-tls-certs\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.399104 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.399128 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04899be-cfe2-4870-ad3c-735148e7bcdd-config-data\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.399170 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.399242 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.399290 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-config-data\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.399972 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d04899be-cfe2-4870-ad3c-735148e7bcdd-logs\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.400335 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32ede37e-d419-4c76-932e-af1de102a9ad-logs\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.408892 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-config-data\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.417848 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04899be-cfe2-4870-ad3c-735148e7bcdd-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.419356 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.419740 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.420702 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.421735 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04899be-cfe2-4870-ad3c-735148e7bcdd-config-data\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.421814 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ede37e-d419-4c76-932e-af1de102a9ad-public-tls-certs\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.428701 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-925bd\" (UniqueName: \"kubernetes.io/projected/d04899be-cfe2-4870-ad3c-735148e7bcdd-kube-api-access-925bd\") pod \"watcher-applier-0\" (UID: \"d04899be-cfe2-4870-ad3c-735148e7bcdd\") " pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.437648 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlff9\" (UniqueName: \"kubernetes.io/projected/32ede37e-d419-4c76-932e-af1de102a9ad-kube-api-access-hlff9\") pod \"watcher-api-0\" (UID: \"32ede37e-d419-4c76-932e-af1de102a9ad\") " pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.467920 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.527448 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.597928 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.885964 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6d769bd5f4-zkxj4" Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.959935 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-754cbc7dbb-wbdt6"] Sep 30 08:22:36 crc kubenswrapper[4810]: I0930 08:22:36.982135 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Sep 30 08:22:37 crc kubenswrapper[4810]: I0930 08:22:37.062739 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"32ede37e-d419-4c76-932e-af1de102a9ad","Type":"ContainerStarted","Data":"f7f78c02e1975d088d174dc52816ab2299c3a3cf3e7a8cd6aad836fea8067490"} Sep 30 08:22:37 crc kubenswrapper[4810]: I0930 08:22:37.064639 4810 generic.go:334] "Generic (PLEG): container finished" podID="ff6ce927-f393-4895-a7b6-af266138c8c1" containerID="a8483f9453c3170461fd7f633d48cb345a6e19d91c2506f88094b43631ddb18c" exitCode=0 Sep 30 08:22:37 crc kubenswrapper[4810]: I0930 08:22:37.064720 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4c6qh" event={"ID":"ff6ce927-f393-4895-a7b6-af266138c8c1","Type":"ContainerDied","Data":"a8483f9453c3170461fd7f633d48cb345a6e19d91c2506f88094b43631ddb18c"} Sep 30 08:22:37 crc kubenswrapper[4810]: I0930 08:22:37.067833 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-z7k2l" event={"ID":"e199d2a8-2a58-47c7-8018-13b29c37c2ad","Type":"ContainerStarted","Data":"60c921d9ce7b936f7c16b702afca10ea57bac1a8f2e2198e37531bdc3102bc8c"} Sep 30 08:22:37 crc kubenswrapper[4810]: I0930 08:22:37.075988 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-754cbc7dbb-wbdt6" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerName="horizon-log" containerID="cri-o://65c41b5d112da8a316221d00dc0b08a61accf182470442b32f5aa0e475f2698f" gracePeriod=30 Sep 30 08:22:37 crc kubenswrapper[4810]: I0930 08:22:37.076426 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-754cbc7dbb-wbdt6" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerName="horizon" containerID="cri-o://a85450da47bf890fcf9ae454a6f57b37faf86b24f277f113255feb929be07b4c" gracePeriod=30 Sep 30 08:22:37 crc kubenswrapper[4810]: I0930 08:22:37.110629 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Sep 30 08:22:37 crc kubenswrapper[4810]: I0930 08:22:37.113785 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-z7k2l" podStartSLOduration=4.3762884 podStartE2EDuration="54.113768034s" podCreationTimestamp="2025-09-30 08:21:43 +0000 UTC" firstStartedPulling="2025-09-30 08:21:45.50168527 +0000 UTC m=+1128.953884537" lastFinishedPulling="2025-09-30 08:22:35.239164904 +0000 UTC m=+1178.691364171" observedRunningTime="2025-09-30 08:22:37.113274069 +0000 UTC m=+1180.565473326" watchObservedRunningTime="2025-09-30 08:22:37.113768034 +0000 UTC m=+1180.565967301" Sep 30 08:22:37 crc kubenswrapper[4810]: I0930 08:22:37.326866 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="277a70ee-91dd-4198-9abd-7e6bb909bec0" path="/var/lib/kubelet/pods/277a70ee-91dd-4198-9abd-7e6bb909bec0/volumes" Sep 30 08:22:37 crc kubenswrapper[4810]: I0930 08:22:37.327640 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acf21e98-01dc-4a20-aa79-bb5549e59f22" path="/var/lib/kubelet/pods/acf21e98-01dc-4a20-aa79-bb5549e59f22/volumes" Sep 30 08:22:37 crc kubenswrapper[4810]: I0930 08:22:37.428620 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.091944 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"d04899be-cfe2-4870-ad3c-735148e7bcdd","Type":"ContainerStarted","Data":"8b8cb98e00f27b0263ad8cc6760e9832205af05afffaca58bfa92d563a4a2ce5"} Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.092223 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"d04899be-cfe2-4870-ad3c-735148e7bcdd","Type":"ContainerStarted","Data":"0ceb18e80a0378cd3858dabe5d27c971830af71fdc40e6343b9d8b4eb9fefec8"} Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.097504 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"32ede37e-d419-4c76-932e-af1de102a9ad","Type":"ContainerStarted","Data":"d051777f852c2a667df1408b6518fbfa616b27283c22b86a36c14a89e30af10a"} Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.097563 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"32ede37e-d419-4c76-932e-af1de102a9ad","Type":"ContainerStarted","Data":"357c99bc6c112745abf15ba32c0e7dc1b57a2b478689b18caf5aaedb775ebcc0"} Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.097714 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.109467 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=2.109449215 podStartE2EDuration="2.109449215s" podCreationTimestamp="2025-09-30 08:22:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:38.108836478 +0000 UTC m=+1181.561035765" watchObservedRunningTime="2025-09-30 08:22:38.109449215 +0000 UTC m=+1181.561648482" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.114346 4810 generic.go:334] "Generic (PLEG): container finished" podID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerID="a85450da47bf890fcf9ae454a6f57b37faf86b24f277f113255feb929be07b4c" exitCode=0 Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.114545 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754cbc7dbb-wbdt6" event={"ID":"c176d78f-1abf-4191-94f8-8ee7be102b18","Type":"ContainerDied","Data":"a85450da47bf890fcf9ae454a6f57b37faf86b24f277f113255feb929be07b4c"} Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.131942 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=2.131924478 podStartE2EDuration="2.131924478s" podCreationTimestamp="2025-09-30 08:22:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:38.129308612 +0000 UTC m=+1181.581507879" watchObservedRunningTime="2025-09-30 08:22:38.131924478 +0000 UTC m=+1181.584123745" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.630363 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4c6qh" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.750204 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-config-data\") pod \"ff6ce927-f393-4895-a7b6-af266138c8c1\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.750353 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-combined-ca-bundle\") pod \"ff6ce927-f393-4895-a7b6-af266138c8c1\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.750441 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-db-sync-config-data\") pod \"ff6ce927-f393-4895-a7b6-af266138c8c1\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.750483 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqpxq\" (UniqueName: \"kubernetes.io/projected/ff6ce927-f393-4895-a7b6-af266138c8c1-kube-api-access-vqpxq\") pod \"ff6ce927-f393-4895-a7b6-af266138c8c1\" (UID: \"ff6ce927-f393-4895-a7b6-af266138c8c1\") " Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.757200 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ff6ce927-f393-4895-a7b6-af266138c8c1" (UID: "ff6ce927-f393-4895-a7b6-af266138c8c1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.762509 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff6ce927-f393-4895-a7b6-af266138c8c1-kube-api-access-vqpxq" (OuterVolumeSpecName: "kube-api-access-vqpxq") pod "ff6ce927-f393-4895-a7b6-af266138c8c1" (UID: "ff6ce927-f393-4895-a7b6-af266138c8c1"). InnerVolumeSpecName "kube-api-access-vqpxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.780465 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff6ce927-f393-4895-a7b6-af266138c8c1" (UID: "ff6ce927-f393-4895-a7b6-af266138c8c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.829615 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-config-data" (OuterVolumeSpecName: "config-data") pod "ff6ce927-f393-4895-a7b6-af266138c8c1" (UID: "ff6ce927-f393-4895-a7b6-af266138c8c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.853225 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.853700 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.853712 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqpxq\" (UniqueName: \"kubernetes.io/projected/ff6ce927-f393-4895-a7b6-af266138c8c1-kube-api-access-vqpxq\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:38 crc kubenswrapper[4810]: I0930 08:22:38.853726 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff6ce927-f393-4895-a7b6-af266138c8c1-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.132534 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4c6qh" event={"ID":"ff6ce927-f393-4895-a7b6-af266138c8c1","Type":"ContainerDied","Data":"0760ebf1f31bcdf28e93b3d14f8865b225951455df4408d3ee32e89338d0e68e"} Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.132583 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0760ebf1f31bcdf28e93b3d14f8865b225951455df4408d3ee32e89338d0e68e" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.132578 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4c6qh" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.134718 4810 generic.go:334] "Generic (PLEG): container finished" podID="913aa7a2-f310-4b1e-a403-2e5f98b339b5" containerID="f0abbb76c1572c27e8547117cdc22146c0ce28960bd84e36efdb24d3e672034c" exitCode=0 Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.134781 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k5dlb" event={"ID":"913aa7a2-f310-4b1e-a403-2e5f98b339b5","Type":"ContainerDied","Data":"f0abbb76c1572c27e8547117cdc22146c0ce28960bd84e36efdb24d3e672034c"} Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.514327 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f5b594445-wnwfj"] Sep 30 08:22:39 crc kubenswrapper[4810]: E0930 08:22:39.514695 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6ce927-f393-4895-a7b6-af266138c8c1" containerName="glance-db-sync" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.514710 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6ce927-f393-4895-a7b6-af266138c8c1" containerName="glance-db-sync" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.514918 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff6ce927-f393-4895-a7b6-af266138c8c1" containerName="glance-db-sync" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.515913 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.539485 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f5b594445-wnwfj"] Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.669887 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-config\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.670154 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-swift-storage-0\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.670207 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2djp2\" (UniqueName: \"kubernetes.io/projected/c2fbcd16-460f-4cef-83aa-10dae3a13050-kube-api-access-2djp2\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.670286 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-sb\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.670335 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-nb\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.670424 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-svc\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.771755 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-svc\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.771804 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-config\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.771822 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-swift-storage-0\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.771878 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2djp2\" (UniqueName: \"kubernetes.io/projected/c2fbcd16-460f-4cef-83aa-10dae3a13050-kube-api-access-2djp2\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.771921 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-sb\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.771968 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-nb\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.773695 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-swift-storage-0\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.775387 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-config\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.775648 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-svc\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.775683 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-sb\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.776593 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-nb\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.797177 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2djp2\" (UniqueName: \"kubernetes.io/projected/c2fbcd16-460f-4cef-83aa-10dae3a13050-kube-api-access-2djp2\") pod \"dnsmasq-dns-7f5b594445-wnwfj\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:39 crc kubenswrapper[4810]: I0930 08:22:39.879451 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.420604 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f5b594445-wnwfj"] Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.438992 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.441331 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.447593 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-wjg2p" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.447755 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.447993 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.453538 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.591305 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-scripts\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.591437 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbpxx\" (UniqueName: \"kubernetes.io/projected/bbc02306-07ec-4a85-95b9-15a4b14b256d-kube-api-access-rbpxx\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.591470 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-config-data\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.591507 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-logs\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.591523 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.591551 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.591576 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.677045 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.678572 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.682764 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.693060 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-scripts\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.693171 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbpxx\" (UniqueName: \"kubernetes.io/projected/bbc02306-07ec-4a85-95b9-15a4b14b256d-kube-api-access-rbpxx\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.693206 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-config-data\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.693243 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-logs\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.693260 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.693311 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.693340 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.694245 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-logs\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.694604 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.694672 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.701862 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-scripts\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.710572 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-config-data\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.721879 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.724671 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.731165 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbpxx\" (UniqueName: \"kubernetes.io/projected/bbc02306-07ec-4a85-95b9-15a4b14b256d-kube-api-access-rbpxx\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.783542 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.796730 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-logs\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.796821 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.796857 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.796896 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf8gh\" (UniqueName: \"kubernetes.io/projected/c545529d-0706-4063-9e3e-75eabc8b9785-kube-api-access-jf8gh\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.796937 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.797003 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.797038 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.898407 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.898669 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.898793 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-logs\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.898916 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.898999 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.899060 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.899108 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.899356 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf8gh\" (UniqueName: \"kubernetes.io/projected/c545529d-0706-4063-9e3e-75eabc8b9785-kube-api-access-jf8gh\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.899469 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.902155 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-logs\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.903843 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.904241 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.915801 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf8gh\" (UniqueName: \"kubernetes.io/projected/c545529d-0706-4063-9e3e-75eabc8b9785-kube-api-access-jf8gh\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.919651 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.933493 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:40 crc kubenswrapper[4810]: I0930 08:22:40.988590 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Sep 30 08:22:41 crc kubenswrapper[4810]: I0930 08:22:41.002571 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 08:22:41 crc kubenswrapper[4810]: I0930 08:22:41.072778 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 08:22:41 crc kubenswrapper[4810]: I0930 08:22:41.469835 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Sep 30 08:22:41 crc kubenswrapper[4810]: I0930 08:22:41.530448 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Sep 30 08:22:42 crc kubenswrapper[4810]: I0930 08:22:42.120784 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:22:42 crc kubenswrapper[4810]: I0930 08:22:42.194954 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:22:42 crc kubenswrapper[4810]: I0930 08:22:42.773085 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-754cbc7dbb-wbdt6" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.163:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.163:8443: connect: connection refused" Sep 30 08:22:43 crc kubenswrapper[4810]: I0930 08:22:43.186109 4810 generic.go:334] "Generic (PLEG): container finished" podID="e199d2a8-2a58-47c7-8018-13b29c37c2ad" containerID="60c921d9ce7b936f7c16b702afca10ea57bac1a8f2e2198e37531bdc3102bc8c" exitCode=0 Sep 30 08:22:43 crc kubenswrapper[4810]: I0930 08:22:43.186560 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-z7k2l" event={"ID":"e199d2a8-2a58-47c7-8018-13b29c37c2ad","Type":"ContainerDied","Data":"60c921d9ce7b936f7c16b702afca10ea57bac1a8f2e2198e37531bdc3102bc8c"} Sep 30 08:22:43 crc kubenswrapper[4810]: I0930 08:22:43.479061 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:22:43 crc kubenswrapper[4810]: I0930 08:22:43.479111 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Sep 30 08:22:43 crc kubenswrapper[4810]: I0930 08:22:43.479901 4810 scope.go:117] "RemoveContainer" containerID="1f7b36decfbc46430a048f59252a01883a015411e53ee22a10670b789efa6fb8" Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.199473 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k5dlb" event={"ID":"913aa7a2-f310-4b1e-a403-2e5f98b339b5","Type":"ContainerDied","Data":"edcc2f23746b8a250d9181834cc31539888489f4a716c9584e4b00ed6860b8d0"} Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.199525 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edcc2f23746b8a250d9181834cc31539888489f4a716c9584e4b00ed6860b8d0" Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.284215 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.389894 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-db-sync-config-data\") pod \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.390000 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-combined-ca-bundle\") pod \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.390028 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdl48\" (UniqueName: \"kubernetes.io/projected/913aa7a2-f310-4b1e-a403-2e5f98b339b5-kube-api-access-fdl48\") pod \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\" (UID: \"913aa7a2-f310-4b1e-a403-2e5f98b339b5\") " Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.397914 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/913aa7a2-f310-4b1e-a403-2e5f98b339b5-kube-api-access-fdl48" (OuterVolumeSpecName: "kube-api-access-fdl48") pod "913aa7a2-f310-4b1e-a403-2e5f98b339b5" (UID: "913aa7a2-f310-4b1e-a403-2e5f98b339b5"). InnerVolumeSpecName "kube-api-access-fdl48". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.398358 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "913aa7a2-f310-4b1e-a403-2e5f98b339b5" (UID: "913aa7a2-f310-4b1e-a403-2e5f98b339b5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.459360 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "913aa7a2-f310-4b1e-a403-2e5f98b339b5" (UID: "913aa7a2-f310-4b1e-a403-2e5f98b339b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.494722 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.494785 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdl48\" (UniqueName: \"kubernetes.io/projected/913aa7a2-f310-4b1e-a403-2e5f98b339b5-kube-api-access-fdl48\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:44 crc kubenswrapper[4810]: I0930 08:22:44.494800 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/913aa7a2-f310-4b1e-a403-2e5f98b339b5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.211010 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k5dlb" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.606231 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5dff5594b-jj7gs"] Sep 30 08:22:45 crc kubenswrapper[4810]: E0930 08:22:45.606839 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913aa7a2-f310-4b1e-a403-2e5f98b339b5" containerName="barbican-db-sync" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.606871 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="913aa7a2-f310-4b1e-a403-2e5f98b339b5" containerName="barbican-db-sync" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.607238 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="913aa7a2-f310-4b1e-a403-2e5f98b339b5" containerName="barbican-db-sync" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.609604 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.611971 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-bfjm2" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.612237 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.612494 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.646317 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-687444b9dc-hxqlq"] Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.647879 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.651679 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.668088 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5dff5594b-jj7gs"] Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.679513 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-687444b9dc-hxqlq"] Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.728546 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff274965-178c-4eff-9841-494634e711f0-config-data-custom\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.728621 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-combined-ca-bundle\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.728676 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff274965-178c-4eff-9841-494634e711f0-combined-ca-bundle\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.728724 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sr9k\" (UniqueName: \"kubernetes.io/projected/ff274965-178c-4eff-9841-494634e711f0-kube-api-access-4sr9k\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.728758 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-config-data-custom\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.728843 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-config-data\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.728894 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff274965-178c-4eff-9841-494634e711f0-logs\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.728941 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff274965-178c-4eff-9841-494634e711f0-config-data\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.729014 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hls5j\" (UniqueName: \"kubernetes.io/projected/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-kube-api-access-hls5j\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.729044 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-logs\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.759612 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f5b594445-wnwfj"] Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.796424 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-574c876857-6t67r"] Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.799413 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.828152 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-574c876857-6t67r"] Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.838230 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff274965-178c-4eff-9841-494634e711f0-config-data-custom\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.838974 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-combined-ca-bundle\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.839050 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff274965-178c-4eff-9841-494634e711f0-combined-ca-bundle\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.839118 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sr9k\" (UniqueName: \"kubernetes.io/projected/ff274965-178c-4eff-9841-494634e711f0-kube-api-access-4sr9k\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.839159 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-config-data-custom\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.839216 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-config-data\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.839251 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff274965-178c-4eff-9841-494634e711f0-logs\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.842801 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff274965-178c-4eff-9841-494634e711f0-logs\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.842941 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff274965-178c-4eff-9841-494634e711f0-config-data\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.843106 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hls5j\" (UniqueName: \"kubernetes.io/projected/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-kube-api-access-hls5j\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.843145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-logs\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.843613 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-logs\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.848534 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-config-data\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.849726 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff274965-178c-4eff-9841-494634e711f0-config-data\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.853974 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff274965-178c-4eff-9841-494634e711f0-combined-ca-bundle\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.859369 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-combined-ca-bundle\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.865826 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hls5j\" (UniqueName: \"kubernetes.io/projected/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-kube-api-access-hls5j\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.867872 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2c5a1aa-c611-4702-a0bb-27693ebb7c7a-config-data-custom\") pod \"barbican-keystone-listener-5dff5594b-jj7gs\" (UID: \"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a\") " pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.868169 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff274965-178c-4eff-9841-494634e711f0-config-data-custom\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.870296 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sr9k\" (UniqueName: \"kubernetes.io/projected/ff274965-178c-4eff-9841-494634e711f0-kube-api-access-4sr9k\") pod \"barbican-worker-687444b9dc-hxqlq\" (UID: \"ff274965-178c-4eff-9841-494634e711f0\") " pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.910937 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-65d8f9c97d-267vl"] Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.914801 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.925561 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.944846 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65d8f9c97d-267vl"] Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.946831 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.947993 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-svc\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.948102 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-swift-storage-0\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.948215 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-nb\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.948257 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-sb\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.948376 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ng7k\" (UniqueName: \"kubernetes.io/projected/55862b13-dfe5-48b0-b4e4-79c38a01b707-kube-api-access-2ng7k\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.948435 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-config\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:45 crc kubenswrapper[4810]: W0930 08:22:45.974496 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2fbcd16_460f_4cef_83aa_10dae3a13050.slice/crio-5c9c7817c2eeae9713adc5efe5ee232b9d84d239bd8701a7c571a91578fb964e WatchSource:0}: Error finding container 5c9c7817c2eeae9713adc5efe5ee232b9d84d239bd8701a7c571a91578fb964e: Status 404 returned error can't find the container with id 5c9c7817c2eeae9713adc5efe5ee232b9d84d239bd8701a7c571a91578fb964e Sep 30 08:22:45 crc kubenswrapper[4810]: I0930 08:22:45.978950 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-687444b9dc-hxqlq" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.050389 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-nb\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.050467 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data-custom\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.050498 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-sb\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.050537 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ng7k\" (UniqueName: \"kubernetes.io/projected/55862b13-dfe5-48b0-b4e4-79c38a01b707-kube-api-access-2ng7k\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.050613 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-config\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.051460 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-svc\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.051592 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-sb\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.052129 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.052375 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-svc\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.052510 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-config\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.052929 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-nb\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.053757 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-swift-storage-0\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.054095 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-combined-ca-bundle\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.054120 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ac68abd-cf5c-4a46-9db1-599d547c3b06-logs\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.054147 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xczhb\" (UniqueName: \"kubernetes.io/projected/5ac68abd-cf5c-4a46-9db1-599d547c3b06-kube-api-access-xczhb\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.054732 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-swift-storage-0\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.070436 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ng7k\" (UniqueName: \"kubernetes.io/projected/55862b13-dfe5-48b0-b4e4-79c38a01b707-kube-api-access-2ng7k\") pod \"dnsmasq-dns-574c876857-6t67r\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.117976 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.130073 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.156905 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-combined-ca-bundle\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.157592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ac68abd-cf5c-4a46-9db1-599d547c3b06-logs\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.157781 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xczhb\" (UniqueName: \"kubernetes.io/projected/5ac68abd-cf5c-4a46-9db1-599d547c3b06-kube-api-access-xczhb\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.158014 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data-custom\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.158232 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ac68abd-cf5c-4a46-9db1-599d547c3b06-logs\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.158527 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.161103 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-combined-ca-bundle\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.162012 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data-custom\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.163670 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.177826 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xczhb\" (UniqueName: \"kubernetes.io/projected/5ac68abd-cf5c-4a46-9db1-599d547c3b06-kube-api-access-xczhb\") pod \"barbican-api-65d8f9c97d-267vl\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.226841 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" event={"ID":"c2fbcd16-460f-4cef-83aa-10dae3a13050","Type":"ContainerStarted","Data":"5c9c7817c2eeae9713adc5efe5ee232b9d84d239bd8701a7c571a91578fb964e"} Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.229178 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-z7k2l" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.229173 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-z7k2l" event={"ID":"e199d2a8-2a58-47c7-8018-13b29c37c2ad","Type":"ContainerDied","Data":"54d060a5e6c7bd24bf9c16899a43f288f357b52be1955c22fec188188163b352"} Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.229454 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54d060a5e6c7bd24bf9c16899a43f288f357b52be1955c22fec188188163b352" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.260000 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-scripts\") pod \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.260338 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qggnm\" (UniqueName: \"kubernetes.io/projected/e199d2a8-2a58-47c7-8018-13b29c37c2ad-kube-api-access-qggnm\") pod \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.260846 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-db-sync-config-data\") pod \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.261090 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e199d2a8-2a58-47c7-8018-13b29c37c2ad-etc-machine-id\") pod \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.261241 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-config-data\") pod \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.261411 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-combined-ca-bundle\") pod \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\" (UID: \"e199d2a8-2a58-47c7-8018-13b29c37c2ad\") " Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.261652 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e199d2a8-2a58-47c7-8018-13b29c37c2ad-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e199d2a8-2a58-47c7-8018-13b29c37c2ad" (UID: "e199d2a8-2a58-47c7-8018-13b29c37c2ad"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.262655 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e199d2a8-2a58-47c7-8018-13b29c37c2ad-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.264352 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e199d2a8-2a58-47c7-8018-13b29c37c2ad" (UID: "e199d2a8-2a58-47c7-8018-13b29c37c2ad"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.264863 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e199d2a8-2a58-47c7-8018-13b29c37c2ad-kube-api-access-qggnm" (OuterVolumeSpecName: "kube-api-access-qggnm") pod "e199d2a8-2a58-47c7-8018-13b29c37c2ad" (UID: "e199d2a8-2a58-47c7-8018-13b29c37c2ad"). InnerVolumeSpecName "kube-api-access-qggnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.265336 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-scripts" (OuterVolumeSpecName: "scripts") pod "e199d2a8-2a58-47c7-8018-13b29c37c2ad" (UID: "e199d2a8-2a58-47c7-8018-13b29c37c2ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.292494 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e199d2a8-2a58-47c7-8018-13b29c37c2ad" (UID: "e199d2a8-2a58-47c7-8018-13b29c37c2ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.333760 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-config-data" (OuterVolumeSpecName: "config-data") pod "e199d2a8-2a58-47c7-8018-13b29c37c2ad" (UID: "e199d2a8-2a58-47c7-8018-13b29c37c2ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.365561 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.365597 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qggnm\" (UniqueName: \"kubernetes.io/projected/e199d2a8-2a58-47c7-8018-13b29c37c2ad-kube-api-access-qggnm\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.365612 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.365623 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.365633 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e199d2a8-2a58-47c7-8018-13b29c37c2ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.422519 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.469407 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.479692 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.528590 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Sep 30 08:22:46 crc kubenswrapper[4810]: I0930 08:22:46.563791 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.289958 4810 generic.go:334] "Generic (PLEG): container finished" podID="dc2cb8f1-4585-43b1-865c-bea9b61ef20c" containerID="02ef2f083fab19cb6ae206c29afed82056cbba7bdb6457c6b4e1f1c10efd387f" exitCode=137 Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.290254 4810 generic.go:334] "Generic (PLEG): container finished" podID="dc2cb8f1-4585-43b1-865c-bea9b61ef20c" containerID="96a6199f09eee930d84db591b76f58776c48a86f7139b00970301945fc37beba" exitCode=137 Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.292695 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-98d79fd7c-kz5wm" event={"ID":"dc2cb8f1-4585-43b1-865c-bea9b61ef20c","Type":"ContainerDied","Data":"02ef2f083fab19cb6ae206c29afed82056cbba7bdb6457c6b4e1f1c10efd387f"} Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.292756 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-98d79fd7c-kz5wm" event={"ID":"dc2cb8f1-4585-43b1-865c-bea9b61ef20c","Type":"ContainerDied","Data":"96a6199f09eee930d84db591b76f58776c48a86f7139b00970301945fc37beba"} Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.300671 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.579351 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.579649 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 08:22:47 crc kubenswrapper[4810]: E0930 08:22:47.579966 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e199d2a8-2a58-47c7-8018-13b29c37c2ad" containerName="cinder-db-sync" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.579979 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e199d2a8-2a58-47c7-8018-13b29c37c2ad" containerName="cinder-db-sync" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.580195 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e199d2a8-2a58-47c7-8018-13b29c37c2ad" containerName="cinder-db-sync" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.581222 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.581238 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-574c876857-6t67r"] Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.581250 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5595b896c9-xltdz"] Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.582346 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.610511 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5595b896c9-xltdz"] Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.610564 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.612441 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.616818 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.624454 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.624751 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.637116 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.637319 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-lkx2n" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.654803 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.654867 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-nb\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.654918 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.654947 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/975d2888-e21f-4ff1-88b4-841427c270e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.654966 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksnxn\" (UniqueName: \"kubernetes.io/projected/975d2888-e21f-4ff1-88b4-841427c270e5-kube-api-access-ksnxn\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.654983 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-config\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.654999 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thk8p\" (UniqueName: \"kubernetes.io/projected/e070c75e-313a-423c-88d7-b4975a8d113d-kube-api-access-thk8p\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.655062 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-swift-storage-0\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.655086 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.655131 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-svc\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.655159 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-sb\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.655182 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.690837 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 30 08:22:47 crc kubenswrapper[4810]: E0930 08:22:47.691634 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2cb8f1-4585-43b1-865c-bea9b61ef20c" containerName="horizon-log" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.691669 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2cb8f1-4585-43b1-865c-bea9b61ef20c" containerName="horizon-log" Sep 30 08:22:47 crc kubenswrapper[4810]: E0930 08:22:47.691692 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2cb8f1-4585-43b1-865c-bea9b61ef20c" containerName="horizon" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.691699 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2cb8f1-4585-43b1-865c-bea9b61ef20c" containerName="horizon" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.692034 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc2cb8f1-4585-43b1-865c-bea9b61ef20c" containerName="horizon-log" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.692274 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc2cb8f1-4585-43b1-865c-bea9b61ef20c" containerName="horizon" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.693883 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.697633 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.704724 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.760432 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-scripts\") pod \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.760752 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-horizon-secret-key\") pod \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.760784 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-config-data\") pod \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.760818 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw47h\" (UniqueName: \"kubernetes.io/projected/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-kube-api-access-hw47h\") pod \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.760868 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-logs\") pod \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\" (UID: \"dc2cb8f1-4585-43b1-865c-bea9b61ef20c\") " Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761483 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761556 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-nb\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761603 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761636 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data-custom\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761668 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/975d2888-e21f-4ff1-88b4-841427c270e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761690 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-scripts\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761710 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksnxn\" (UniqueName: \"kubernetes.io/projected/975d2888-e21f-4ff1-88b4-841427c270e5-kube-api-access-ksnxn\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761750 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-config\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761778 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thk8p\" (UniqueName: \"kubernetes.io/projected/e070c75e-313a-423c-88d7-b4975a8d113d-kube-api-access-thk8p\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761805 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761830 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761866 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-swift-storage-0\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761899 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761927 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761962 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-svc\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.761987 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-sb\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.762009 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.762030 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgb2w\" (UniqueName: \"kubernetes.io/projected/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-kube-api-access-pgb2w\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.762050 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-logs\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.763605 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-config\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.765610 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-svc\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.769659 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/975d2888-e21f-4ff1-88b4-841427c270e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.770147 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-logs" (OuterVolumeSpecName: "logs") pod "dc2cb8f1-4585-43b1-865c-bea9b61ef20c" (UID: "dc2cb8f1-4585-43b1-865c-bea9b61ef20c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.776483 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-sb\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.776524 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.777183 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-nb\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.783068 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-swift-storage-0\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.786103 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "dc2cb8f1-4585-43b1-865c-bea9b61ef20c" (UID: "dc2cb8f1-4585-43b1-865c-bea9b61ef20c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.785723 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.797343 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.804087 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.804125 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksnxn\" (UniqueName: \"kubernetes.io/projected/975d2888-e21f-4ff1-88b4-841427c270e5-kube-api-access-ksnxn\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.805213 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.818606 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-kube-api-access-hw47h" (OuterVolumeSpecName: "kube-api-access-hw47h") pod "dc2cb8f1-4585-43b1-865c-bea9b61ef20c" (UID: "dc2cb8f1-4585-43b1-865c-bea9b61ef20c"). InnerVolumeSpecName "kube-api-access-hw47h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.828667 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thk8p\" (UniqueName: \"kubernetes.io/projected/e070c75e-313a-423c-88d7-b4975a8d113d-kube-api-access-thk8p\") pod \"dnsmasq-dns-5595b896c9-xltdz\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:47 crc kubenswrapper[4810]: E0930 08:22:47.829653 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"sg-core\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.833695 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-scripts" (OuterVolumeSpecName: "scripts") pod "dc2cb8f1-4585-43b1-865c-bea9b61ef20c" (UID: "dc2cb8f1-4585-43b1-865c-bea9b61ef20c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.845720 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-config-data" (OuterVolumeSpecName: "config-data") pod "dc2cb8f1-4585-43b1-865c-bea9b61ef20c" (UID: "dc2cb8f1-4585-43b1-865c-bea9b61ef20c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.875930 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data-custom\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876003 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-scripts\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876063 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876092 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876193 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876288 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgb2w\" (UniqueName: \"kubernetes.io/projected/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-kube-api-access-pgb2w\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876312 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-logs\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876482 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876494 4810 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876506 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876515 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw47h\" (UniqueName: \"kubernetes.io/projected/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-kube-api-access-hw47h\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876527 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc2cb8f1-4585-43b1-865c-bea9b61ef20c-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876886 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-logs\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.876937 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.884658 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.886207 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data-custom\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.887979 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-scripts\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.889694 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.905192 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgb2w\" (UniqueName: \"kubernetes.io/projected/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-kube-api-access-pgb2w\") pod \"cinder-api-0\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " pod="openstack/cinder-api-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.964057 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 08:22:47 crc kubenswrapper[4810]: I0930 08:22:47.985955 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.023939 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.278909 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-574c876857-6t67r"] Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.320479 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-687444b9dc-hxqlq"] Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.324510 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92bcbff2-f0a3-4303-96cc-5f3177731f6a","Type":"ContainerStarted","Data":"def3adc74d1504e6cf43b5e58fe58dc2fbbf5e1b517086401891309b91ec588c"} Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.324692 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" containerName="ceilometer-notification-agent" containerID="cri-o://fa6601386e8675aa2cb5e215eff6c3f2aa637b95279aa81ba73c2faf6c08b6f4" gracePeriod=30 Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.324777 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.325026 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" containerName="proxy-httpd" containerID="cri-o://def3adc74d1504e6cf43b5e58fe58dc2fbbf5e1b517086401891309b91ec588c" gracePeriod=30 Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.328886 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c545529d-0706-4063-9e3e-75eabc8b9785","Type":"ContainerStarted","Data":"467b0933e622471a2dc09057f36c318aa5122a659511a9a2822af2149552dca7"} Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.329416 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5dff5594b-jj7gs"] Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.348946 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerStarted","Data":"97265a9c186e6460557751232f731c5bc9eac539bb43e0580b1ec02387f960fc"} Sep 30 08:22:48 crc kubenswrapper[4810]: E0930 08:22:48.382175 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2fbcd16_460f_4cef_83aa_10dae3a13050.slice/crio-conmon-e7e185e2439534208d918105d0d827e8ff9a42a23624373469101ec0b1111e53.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2fbcd16_460f_4cef_83aa_10dae3a13050.slice/crio-e7e185e2439534208d918105d0d827e8ff9a42a23624373469101ec0b1111e53.scope\": RecentStats: unable to find data in memory cache]" Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.385053 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-98d79fd7c-kz5wm" Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.386679 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-98d79fd7c-kz5wm" event={"ID":"dc2cb8f1-4585-43b1-865c-bea9b61ef20c","Type":"ContainerDied","Data":"35e373788398a18fee8bdf3c985ae464a5452dca2be0c15ccca9ae00f76b7502"} Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.386720 4810 scope.go:117] "RemoveContainer" containerID="02ef2f083fab19cb6ae206c29afed82056cbba7bdb6457c6b4e1f1c10efd387f" Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.399799 4810 generic.go:334] "Generic (PLEG): container finished" podID="c2fbcd16-460f-4cef-83aa-10dae3a13050" containerID="e7e185e2439534208d918105d0d827e8ff9a42a23624373469101ec0b1111e53" exitCode=0 Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.399876 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" event={"ID":"c2fbcd16-460f-4cef-83aa-10dae3a13050","Type":"ContainerDied","Data":"e7e185e2439534208d918105d0d827e8ff9a42a23624373469101ec0b1111e53"} Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.484462 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-98d79fd7c-kz5wm"] Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.508535 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-98d79fd7c-kz5wm"] Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.575233 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65d8f9c97d-267vl"] Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.712123 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.716136 4810 scope.go:117] "RemoveContainer" containerID="96a6199f09eee930d84db591b76f58776c48a86f7139b00970301945fc37beba" Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.779422 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5595b896c9-xltdz"] Sep 30 08:22:48 crc kubenswrapper[4810]: I0930 08:22:48.807309 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.089637 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.370660 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc2cb8f1-4585-43b1-865c-bea9b61ef20c" path="/var/lib/kubelet/pods/dc2cb8f1-4585-43b1-865c-bea9b61ef20c/volumes" Sep 30 08:22:49 crc kubenswrapper[4810]: W0930 08:22:49.370766 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79ec9ff9_6c95_4fa1_86d4_e6e028b8e02b.slice/crio-d9fba371e8c661458905bde13739c6e7f14605eb84de0e64580b79ff49ee28d6 WatchSource:0}: Error finding container d9fba371e8c661458905bde13739c6e7f14605eb84de0e64580b79ff49ee28d6: Status 404 returned error can't find the container with id d9fba371e8c661458905bde13739c6e7f14605eb84de0e64580b79ff49ee28d6 Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.422740 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b","Type":"ContainerStarted","Data":"d9fba371e8c661458905bde13739c6e7f14605eb84de0e64580b79ff49ee28d6"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.428451 4810 generic.go:334] "Generic (PLEG): container finished" podID="55862b13-dfe5-48b0-b4e4-79c38a01b707" containerID="14a37d695cb94b01cc26ab257ce43c96141d7c47cdd5f0bb1a6e4b2c8e8af167" exitCode=0 Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.428522 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574c876857-6t67r" event={"ID":"55862b13-dfe5-48b0-b4e4-79c38a01b707","Type":"ContainerDied","Data":"14a37d695cb94b01cc26ab257ce43c96141d7c47cdd5f0bb1a6e4b2c8e8af167"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.428547 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574c876857-6t67r" event={"ID":"55862b13-dfe5-48b0-b4e4-79c38a01b707","Type":"ContainerStarted","Data":"8935e20c47bbff4574a3f91e2385549ca93b3dddc2aca4795c4de6a3f0c5d7b9"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.461450 4810 generic.go:334] "Generic (PLEG): container finished" podID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" containerID="def3adc74d1504e6cf43b5e58fe58dc2fbbf5e1b517086401891309b91ec588c" exitCode=0 Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.461484 4810 generic.go:334] "Generic (PLEG): container finished" podID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" containerID="fa6601386e8675aa2cb5e215eff6c3f2aa637b95279aa81ba73c2faf6c08b6f4" exitCode=0 Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.461530 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92bcbff2-f0a3-4303-96cc-5f3177731f6a","Type":"ContainerDied","Data":"def3adc74d1504e6cf43b5e58fe58dc2fbbf5e1b517086401891309b91ec588c"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.461558 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92bcbff2-f0a3-4303-96cc-5f3177731f6a","Type":"ContainerDied","Data":"fa6601386e8675aa2cb5e215eff6c3f2aa637b95279aa81ba73c2faf6c08b6f4"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.470128 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" event={"ID":"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a","Type":"ContainerStarted","Data":"c47df0f78b9bdb186011e1d4318a92cc477cb717511b83a34939bea12038e6cc"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.473397 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c545529d-0706-4063-9e3e-75eabc8b9785","Type":"ContainerStarted","Data":"60451eb9e86ba4154b35b83174bea31f0296f5d0cfeb85eaf93a33f09ba951f1"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.474590 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-687444b9dc-hxqlq" event={"ID":"ff274965-178c-4eff-9841-494634e711f0","Type":"ContainerStarted","Data":"a1d4edd36d15347d935aff3949ff82524f4dc09dc23d2d3b71e5903aa6d9127c"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.476948 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bbc02306-07ec-4a85-95b9-15a4b14b256d","Type":"ContainerStarted","Data":"6d65c47b8d1307d0ae847fa6237a67377342a1f317c6c57e14b086b0ff86c91a"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.478200 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" event={"ID":"c2fbcd16-460f-4cef-83aa-10dae3a13050","Type":"ContainerDied","Data":"5c9c7817c2eeae9713adc5efe5ee232b9d84d239bd8701a7c571a91578fb964e"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.478218 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c9c7817c2eeae9713adc5efe5ee232b9d84d239bd8701a7c571a91578fb964e" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.483509 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"975d2888-e21f-4ff1-88b4-841427c270e5","Type":"ContainerStarted","Data":"842fba3d2eb546de17ec1de7455e43a4d477fc885a73207d870531d8c4c1f7ea"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.536377 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65d8f9c97d-267vl" event={"ID":"5ac68abd-cf5c-4a46-9db1-599d547c3b06","Type":"ContainerStarted","Data":"cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.536430 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65d8f9c97d-267vl" event={"ID":"5ac68abd-cf5c-4a46-9db1-599d547c3b06","Type":"ContainerStarted","Data":"730c6d8f1bd112a78dda99f986a7691880de092ca2d81faef317b0f3a0882db4"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.552469 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" event={"ID":"e070c75e-313a-423c-88d7-b4975a8d113d","Type":"ContainerStarted","Data":"febc3f449b977933be184110ef07d9be808e55cd56857f2b3e523f9c4ece6a32"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.552517 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" event={"ID":"e070c75e-313a-423c-88d7-b4975a8d113d","Type":"ContainerStarted","Data":"3f0eb234189f0878328e8e05900fb638f809f20471713eb6b79150ff96ddebf1"} Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.567255 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.662067 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-sb\") pod \"c2fbcd16-460f-4cef-83aa-10dae3a13050\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.662545 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-nb\") pod \"c2fbcd16-460f-4cef-83aa-10dae3a13050\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.662576 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-config\") pod \"c2fbcd16-460f-4cef-83aa-10dae3a13050\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.662616 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-swift-storage-0\") pod \"c2fbcd16-460f-4cef-83aa-10dae3a13050\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.662645 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2djp2\" (UniqueName: \"kubernetes.io/projected/c2fbcd16-460f-4cef-83aa-10dae3a13050-kube-api-access-2djp2\") pod \"c2fbcd16-460f-4cef-83aa-10dae3a13050\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.662690 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-svc\") pod \"c2fbcd16-460f-4cef-83aa-10dae3a13050\" (UID: \"c2fbcd16-460f-4cef-83aa-10dae3a13050\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.684157 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2fbcd16-460f-4cef-83aa-10dae3a13050-kube-api-access-2djp2" (OuterVolumeSpecName: "kube-api-access-2djp2") pod "c2fbcd16-460f-4cef-83aa-10dae3a13050" (UID: "c2fbcd16-460f-4cef-83aa-10dae3a13050"). InnerVolumeSpecName "kube-api-access-2djp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.701185 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c2fbcd16-460f-4cef-83aa-10dae3a13050" (UID: "c2fbcd16-460f-4cef-83aa-10dae3a13050"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.722561 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c2fbcd16-460f-4cef-83aa-10dae3a13050" (UID: "c2fbcd16-460f-4cef-83aa-10dae3a13050"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.732398 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c2fbcd16-460f-4cef-83aa-10dae3a13050" (UID: "c2fbcd16-460f-4cef-83aa-10dae3a13050"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.737512 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c2fbcd16-460f-4cef-83aa-10dae3a13050" (UID: "c2fbcd16-460f-4cef-83aa-10dae3a13050"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.764915 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.764944 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.764956 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2djp2\" (UniqueName: \"kubernetes.io/projected/c2fbcd16-460f-4cef-83aa-10dae3a13050-kube-api-access-2djp2\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.764966 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.764976 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.801567 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.816119 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-config" (OuterVolumeSpecName: "config") pod "c2fbcd16-460f-4cef-83aa-10dae3a13050" (UID: "c2fbcd16-460f-4cef-83aa-10dae3a13050"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.866375 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-combined-ca-bundle\") pod \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.866484 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmx6l\" (UniqueName: \"kubernetes.io/projected/92bcbff2-f0a3-4303-96cc-5f3177731f6a-kube-api-access-qmx6l\") pod \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.866536 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-config-data\") pod \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.866574 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-run-httpd\") pod \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.866629 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-scripts\") pod \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.866683 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-log-httpd\") pod \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.866793 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-sg-core-conf-yaml\") pod \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\" (UID: \"92bcbff2-f0a3-4303-96cc-5f3177731f6a\") " Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.867416 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2fbcd16-460f-4cef-83aa-10dae3a13050-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.867938 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "92bcbff2-f0a3-4303-96cc-5f3177731f6a" (UID: "92bcbff2-f0a3-4303-96cc-5f3177731f6a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.868873 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "92bcbff2-f0a3-4303-96cc-5f3177731f6a" (UID: "92bcbff2-f0a3-4303-96cc-5f3177731f6a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.874131 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "92bcbff2-f0a3-4303-96cc-5f3177731f6a" (UID: "92bcbff2-f0a3-4303-96cc-5f3177731f6a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.880463 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92bcbff2-f0a3-4303-96cc-5f3177731f6a-kube-api-access-qmx6l" (OuterVolumeSpecName: "kube-api-access-qmx6l") pod "92bcbff2-f0a3-4303-96cc-5f3177731f6a" (UID: "92bcbff2-f0a3-4303-96cc-5f3177731f6a"). InnerVolumeSpecName "kube-api-access-qmx6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.924560 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-scripts" (OuterVolumeSpecName: "scripts") pod "92bcbff2-f0a3-4303-96cc-5f3177731f6a" (UID: "92bcbff2-f0a3-4303-96cc-5f3177731f6a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.969556 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.969586 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.969596 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.969606 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmx6l\" (UniqueName: \"kubernetes.io/projected/92bcbff2-f0a3-4303-96cc-5f3177731f6a-kube-api-access-qmx6l\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:49 crc kubenswrapper[4810]: I0930 08:22:49.969618 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92bcbff2-f0a3-4303-96cc-5f3177731f6a-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.106369 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-config-data" (OuterVolumeSpecName: "config-data") pod "92bcbff2-f0a3-4303-96cc-5f3177731f6a" (UID: "92bcbff2-f0a3-4303-96cc-5f3177731f6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.114533 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92bcbff2-f0a3-4303-96cc-5f3177731f6a" (UID: "92bcbff2-f0a3-4303-96cc-5f3177731f6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.175608 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.175654 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92bcbff2-f0a3-4303-96cc-5f3177731f6a-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.220169 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.277368 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ng7k\" (UniqueName: \"kubernetes.io/projected/55862b13-dfe5-48b0-b4e4-79c38a01b707-kube-api-access-2ng7k\") pod \"55862b13-dfe5-48b0-b4e4-79c38a01b707\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.277423 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-nb\") pod \"55862b13-dfe5-48b0-b4e4-79c38a01b707\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.277522 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-sb\") pod \"55862b13-dfe5-48b0-b4e4-79c38a01b707\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.277594 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-config\") pod \"55862b13-dfe5-48b0-b4e4-79c38a01b707\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.277648 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-swift-storage-0\") pod \"55862b13-dfe5-48b0-b4e4-79c38a01b707\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.277691 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-svc\") pod \"55862b13-dfe5-48b0-b4e4-79c38a01b707\" (UID: \"55862b13-dfe5-48b0-b4e4-79c38a01b707\") " Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.298778 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55862b13-dfe5-48b0-b4e4-79c38a01b707-kube-api-access-2ng7k" (OuterVolumeSpecName: "kube-api-access-2ng7k") pod "55862b13-dfe5-48b0-b4e4-79c38a01b707" (UID: "55862b13-dfe5-48b0-b4e4-79c38a01b707"). InnerVolumeSpecName "kube-api-access-2ng7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.369012 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "55862b13-dfe5-48b0-b4e4-79c38a01b707" (UID: "55862b13-dfe5-48b0-b4e4-79c38a01b707"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.380736 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ng7k\" (UniqueName: \"kubernetes.io/projected/55862b13-dfe5-48b0-b4e4-79c38a01b707-kube-api-access-2ng7k\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.380795 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.416854 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "55862b13-dfe5-48b0-b4e4-79c38a01b707" (UID: "55862b13-dfe5-48b0-b4e4-79c38a01b707"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.465819 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "55862b13-dfe5-48b0-b4e4-79c38a01b707" (UID: "55862b13-dfe5-48b0-b4e4-79c38a01b707"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.466345 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "55862b13-dfe5-48b0-b4e4-79c38a01b707" (UID: "55862b13-dfe5-48b0-b4e4-79c38a01b707"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.474491 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-config" (OuterVolumeSpecName: "config") pod "55862b13-dfe5-48b0-b4e4-79c38a01b707" (UID: "55862b13-dfe5-48b0-b4e4-79c38a01b707"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.482508 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.482544 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.482556 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.482565 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55862b13-dfe5-48b0-b4e4-79c38a01b707-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.612741 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65d8f9c97d-267vl" event={"ID":"5ac68abd-cf5c-4a46-9db1-599d547c3b06","Type":"ContainerStarted","Data":"fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351"} Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.613396 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.613436 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.628757 4810 generic.go:334] "Generic (PLEG): container finished" podID="e070c75e-313a-423c-88d7-b4975a8d113d" containerID="febc3f449b977933be184110ef07d9be808e55cd56857f2b3e523f9c4ece6a32" exitCode=0 Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.629088 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" event={"ID":"e070c75e-313a-423c-88d7-b4975a8d113d","Type":"ContainerDied","Data":"febc3f449b977933be184110ef07d9be808e55cd56857f2b3e523f9c4ece6a32"} Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.629184 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" event={"ID":"e070c75e-313a-423c-88d7-b4975a8d113d","Type":"ContainerStarted","Data":"1adc7727d0e20d062781ac5384801b5627a66f65be588683002008197f7425dd"} Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.630064 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.643722 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-65d8f9c97d-267vl" podStartSLOduration=5.643702031 podStartE2EDuration="5.643702031s" podCreationTimestamp="2025-09-30 08:22:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:50.630044884 +0000 UTC m=+1194.082244151" watchObservedRunningTime="2025-09-30 08:22:50.643702031 +0000 UTC m=+1194.095901298" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.653391 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bbc02306-07ec-4a85-95b9-15a4b14b256d","Type":"ContainerStarted","Data":"ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd"} Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.654713 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" podStartSLOduration=3.65469752 podStartE2EDuration="3.65469752s" podCreationTimestamp="2025-09-30 08:22:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:50.653340751 +0000 UTC m=+1194.105540018" watchObservedRunningTime="2025-09-30 08:22:50.65469752 +0000 UTC m=+1194.106896787" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.663912 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574c876857-6t67r" event={"ID":"55862b13-dfe5-48b0-b4e4-79c38a01b707","Type":"ContainerDied","Data":"8935e20c47bbff4574a3f91e2385549ca93b3dddc2aca4795c4de6a3f0c5d7b9"} Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.663984 4810 scope.go:117] "RemoveContainer" containerID="14a37d695cb94b01cc26ab257ce43c96141d7c47cdd5f0bb1a6e4b2c8e8af167" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.664040 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574c876857-6t67r" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.672047 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92bcbff2-f0a3-4303-96cc-5f3177731f6a","Type":"ContainerDied","Data":"13b542e536c8600b9a6c7206413ee3fc3f8b1032c6623429f867b1108a6de950"} Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.672142 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.685209 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c545529d-0706-4063-9e3e-75eabc8b9785","Type":"ContainerStarted","Data":"09c06cc5c51068b211fe9ae3b2a140ff729dfb45c98e9055d449f9e357949459"} Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.685288 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c545529d-0706-4063-9e3e-75eabc8b9785" containerName="glance-log" containerID="cri-o://60451eb9e86ba4154b35b83174bea31f0296f5d0cfeb85eaf93a33f09ba951f1" gracePeriod=30 Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.685342 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c545529d-0706-4063-9e3e-75eabc8b9785" containerName="glance-httpd" containerID="cri-o://09c06cc5c51068b211fe9ae3b2a140ff729dfb45c98e9055d449f9e357949459" gracePeriod=30 Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.685405 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f5b594445-wnwfj" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.730799 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-574c876857-6t67r"] Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.743996 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-574c876857-6t67r"] Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.751864 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.751840313 podStartE2EDuration="11.751840313s" podCreationTimestamp="2025-09-30 08:22:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:50.736101216 +0000 UTC m=+1194.188300483" watchObservedRunningTime="2025-09-30 08:22:50.751840313 +0000 UTC m=+1194.204039580" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.808993 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.829504 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.843549 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:22:50 crc kubenswrapper[4810]: E0930 08:22:50.843984 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" containerName="ceilometer-notification-agent" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.844002 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" containerName="ceilometer-notification-agent" Sep 30 08:22:50 crc kubenswrapper[4810]: E0930 08:22:50.844043 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" containerName="proxy-httpd" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.844050 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" containerName="proxy-httpd" Sep 30 08:22:50 crc kubenswrapper[4810]: E0930 08:22:50.844062 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55862b13-dfe5-48b0-b4e4-79c38a01b707" containerName="init" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.844068 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="55862b13-dfe5-48b0-b4e4-79c38a01b707" containerName="init" Sep 30 08:22:50 crc kubenswrapper[4810]: E0930 08:22:50.844080 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2fbcd16-460f-4cef-83aa-10dae3a13050" containerName="init" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.844086 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2fbcd16-460f-4cef-83aa-10dae3a13050" containerName="init" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.844495 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="55862b13-dfe5-48b0-b4e4-79c38a01b707" containerName="init" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.844512 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2fbcd16-460f-4cef-83aa-10dae3a13050" containerName="init" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.844526 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" containerName="ceilometer-notification-agent" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.844539 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" containerName="proxy-httpd" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.849931 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.853470 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.853678 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.856621 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.893539 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-config-data\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.893607 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.893667 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-scripts\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.893699 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-run-httpd\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.893721 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.893772 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgg52\" (UniqueName: \"kubernetes.io/projected/b753b118-389f-432d-a3a8-366ebb8b440d-kube-api-access-kgg52\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.893802 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-log-httpd\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.948947 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f5b594445-wnwfj"] Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.964875 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f5b594445-wnwfj"] Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.995163 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgg52\" (UniqueName: \"kubernetes.io/projected/b753b118-389f-432d-a3a8-366ebb8b440d-kube-api-access-kgg52\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.995236 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-log-httpd\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.995386 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-config-data\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.995427 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.995498 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-scripts\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.995528 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-run-httpd\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.995552 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.996466 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-run-httpd\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:50 crc kubenswrapper[4810]: I0930 08:22:50.996833 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-log-httpd\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.002514 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-config-data\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.021742 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-scripts\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.022082 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.022777 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.024639 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgg52\" (UniqueName: \"kubernetes.io/projected/b753b118-389f-432d-a3a8-366ebb8b440d-kube-api-access-kgg52\") pod \"ceilometer-0\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " pod="openstack/ceilometer-0" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.077229 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5f769c46d7-mgh2q" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.137907 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d889fdd6d-xxkrj"] Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.138385 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d889fdd6d-xxkrj" podUID="8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" containerName="neutron-api" containerID="cri-o://d0eeafa4714463a65c8685739ff271f4d79c5316b1d5cc37ad914db6d677ccab" gracePeriod=30 Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.138871 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d889fdd6d-xxkrj" podUID="8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" containerName="neutron-httpd" containerID="cri-o://21d2347bf42c6eb44e7ee7a5de628b64b5a98e56ec1865f17862029f6ca75cc7" gracePeriod=30 Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.247216 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.329553 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55862b13-dfe5-48b0-b4e4-79c38a01b707" path="/var/lib/kubelet/pods/55862b13-dfe5-48b0-b4e4-79c38a01b707/volumes" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.330654 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92bcbff2-f0a3-4303-96cc-5f3177731f6a" path="/var/lib/kubelet/pods/92bcbff2-f0a3-4303-96cc-5f3177731f6a/volumes" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.331202 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2fbcd16-460f-4cef-83aa-10dae3a13050" path="/var/lib/kubelet/pods/c2fbcd16-460f-4cef-83aa-10dae3a13050/volumes" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.714831 4810 generic.go:334] "Generic (PLEG): container finished" podID="8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" containerID="21d2347bf42c6eb44e7ee7a5de628b64b5a98e56ec1865f17862029f6ca75cc7" exitCode=0 Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.714903 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d889fdd6d-xxkrj" event={"ID":"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523","Type":"ContainerDied","Data":"21d2347bf42c6eb44e7ee7a5de628b64b5a98e56ec1865f17862029f6ca75cc7"} Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.737493 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b","Type":"ContainerStarted","Data":"5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348"} Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.762556 4810 scope.go:117] "RemoveContainer" containerID="def3adc74d1504e6cf43b5e58fe58dc2fbbf5e1b517086401891309b91ec588c" Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.780610 4810 generic.go:334] "Generic (PLEG): container finished" podID="c545529d-0706-4063-9e3e-75eabc8b9785" containerID="09c06cc5c51068b211fe9ae3b2a140ff729dfb45c98e9055d449f9e357949459" exitCode=143 Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.780644 4810 generic.go:334] "Generic (PLEG): container finished" podID="c545529d-0706-4063-9e3e-75eabc8b9785" containerID="60451eb9e86ba4154b35b83174bea31f0296f5d0cfeb85eaf93a33f09ba951f1" exitCode=143 Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.781723 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c545529d-0706-4063-9e3e-75eabc8b9785","Type":"ContainerDied","Data":"09c06cc5c51068b211fe9ae3b2a140ff729dfb45c98e9055d449f9e357949459"} Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.781750 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c545529d-0706-4063-9e3e-75eabc8b9785","Type":"ContainerDied","Data":"60451eb9e86ba4154b35b83174bea31f0296f5d0cfeb85eaf93a33f09ba951f1"} Sep 30 08:22:51 crc kubenswrapper[4810]: I0930 08:22:51.943598 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.596803 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-c6d89f69d-k9g44"] Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.598573 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.603932 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.603982 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.642511 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c6d89f69d-k9g44"] Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.642595 4810 scope.go:117] "RemoveContainer" containerID="fa6601386e8675aa2cb5e215eff6c3f2aa637b95279aa81ba73c2faf6c08b6f4" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.654582 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-config-data-custom\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.654631 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-public-tls-certs\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.654651 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-config-data\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.654717 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-internal-tls-certs\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.655164 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-combined-ca-bundle\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.655404 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcq6k\" (UniqueName: \"kubernetes.io/projected/23454188-6622-4e7c-90f0-2fddf031738e-kube-api-access-wcq6k\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.655446 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23454188-6622-4e7c-90f0-2fddf031738e-logs\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.758296 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-config-data-custom\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.758343 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-public-tls-certs\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.758360 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-config-data\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.758379 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-internal-tls-certs\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.758427 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-combined-ca-bundle\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.758485 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcq6k\" (UniqueName: \"kubernetes.io/projected/23454188-6622-4e7c-90f0-2fddf031738e-kube-api-access-wcq6k\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.758503 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23454188-6622-4e7c-90f0-2fddf031738e-logs\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.759175 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23454188-6622-4e7c-90f0-2fddf031738e-logs\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.764818 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-internal-tls-certs\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.766671 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-config-data\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.774579 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-754cbc7dbb-wbdt6" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.163:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.163:8443: connect: connection refused" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.778711 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-config-data-custom\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.778955 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-combined-ca-bundle\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.779125 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23454188-6622-4e7c-90f0-2fddf031738e-public-tls-certs\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.782377 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcq6k\" (UniqueName: \"kubernetes.io/projected/23454188-6622-4e7c-90f0-2fddf031738e-kube-api-access-wcq6k\") pod \"barbican-api-c6d89f69d-k9g44\" (UID: \"23454188-6622-4e7c-90f0-2fddf031738e\") " pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.795151 4810 generic.go:334] "Generic (PLEG): container finished" podID="8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" containerID="d0eeafa4714463a65c8685739ff271f4d79c5316b1d5cc37ad914db6d677ccab" exitCode=0 Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.795220 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d889fdd6d-xxkrj" event={"ID":"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523","Type":"ContainerDied","Data":"d0eeafa4714463a65c8685739ff271f4d79c5316b1d5cc37ad914db6d677ccab"} Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.804531 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c545529d-0706-4063-9e3e-75eabc8b9785","Type":"ContainerDied","Data":"467b0933e622471a2dc09057f36c318aa5122a659511a9a2822af2149552dca7"} Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.804571 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="467b0933e622471a2dc09057f36c318aa5122a659511a9a2822af2149552dca7" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.805900 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"975d2888-e21f-4ff1-88b4-841427c270e5","Type":"ContainerStarted","Data":"99541a51d283a02a348a3ddbbbe4d6b66ebcdec80efa338931ab2ff8c1db9866"} Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.817111 4810 generic.go:334] "Generic (PLEG): container finished" podID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerID="97265a9c186e6460557751232f731c5bc9eac539bb43e0580b1ec02387f960fc" exitCode=1 Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.818468 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerDied","Data":"97265a9c186e6460557751232f731c5bc9eac539bb43e0580b1ec02387f960fc"} Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.819482 4810 scope.go:117] "RemoveContainer" containerID="97265a9c186e6460557751232f731c5bc9eac539bb43e0580b1ec02387f960fc" Sep 30 08:22:52 crc kubenswrapper[4810]: E0930 08:22:52.819709 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.866867 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.870581 4810 scope.go:117] "RemoveContainer" containerID="1f7b36decfbc46430a048f59252a01883a015411e53ee22a10670b789efa6fb8" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.894932 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.962543 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"c545529d-0706-4063-9e3e-75eabc8b9785\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.962620 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-logs\") pod \"c545529d-0706-4063-9e3e-75eabc8b9785\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.962704 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-config-data\") pod \"c545529d-0706-4063-9e3e-75eabc8b9785\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.962730 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf8gh\" (UniqueName: \"kubernetes.io/projected/c545529d-0706-4063-9e3e-75eabc8b9785-kube-api-access-jf8gh\") pod \"c545529d-0706-4063-9e3e-75eabc8b9785\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.962753 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-combined-ca-bundle\") pod \"c545529d-0706-4063-9e3e-75eabc8b9785\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.962777 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-httpd-run\") pod \"c545529d-0706-4063-9e3e-75eabc8b9785\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.962835 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-scripts\") pod \"c545529d-0706-4063-9e3e-75eabc8b9785\" (UID: \"c545529d-0706-4063-9e3e-75eabc8b9785\") " Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.967962 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-logs" (OuterVolumeSpecName: "logs") pod "c545529d-0706-4063-9e3e-75eabc8b9785" (UID: "c545529d-0706-4063-9e3e-75eabc8b9785"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.968617 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c545529d-0706-4063-9e3e-75eabc8b9785" (UID: "c545529d-0706-4063-9e3e-75eabc8b9785"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.968792 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "c545529d-0706-4063-9e3e-75eabc8b9785" (UID: "c545529d-0706-4063-9e3e-75eabc8b9785"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.979583 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c545529d-0706-4063-9e3e-75eabc8b9785-kube-api-access-jf8gh" (OuterVolumeSpecName: "kube-api-access-jf8gh") pod "c545529d-0706-4063-9e3e-75eabc8b9785" (UID: "c545529d-0706-4063-9e3e-75eabc8b9785"). InnerVolumeSpecName "kube-api-access-jf8gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:52 crc kubenswrapper[4810]: I0930 08:22:52.981698 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-scripts" (OuterVolumeSpecName: "scripts") pod "c545529d-0706-4063-9e3e-75eabc8b9785" (UID: "c545529d-0706-4063-9e3e-75eabc8b9785"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.036438 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-config-data" (OuterVolumeSpecName: "config-data") pod "c545529d-0706-4063-9e3e-75eabc8b9785" (UID: "c545529d-0706-4063-9e3e-75eabc8b9785"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.039111 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c545529d-0706-4063-9e3e-75eabc8b9785" (UID: "c545529d-0706-4063-9e3e-75eabc8b9785"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.064987 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.065026 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf8gh\" (UniqueName: \"kubernetes.io/projected/c545529d-0706-4063-9e3e-75eabc8b9785-kube-api-access-jf8gh\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.065041 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.065050 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.065060 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c545529d-0706-4063-9e3e-75eabc8b9785-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.065105 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.065114 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c545529d-0706-4063-9e3e-75eabc8b9785-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.091751 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.167154 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.412483 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.475087 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-httpd-config\") pod \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.475186 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmndz\" (UniqueName: \"kubernetes.io/projected/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-kube-api-access-zmndz\") pod \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.475226 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-combined-ca-bundle\") pod \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.475345 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-ovndb-tls-certs\") pod \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.475436 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-config\") pod \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\" (UID: \"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523\") " Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.478207 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.478320 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.485551 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-kube-api-access-zmndz" (OuterVolumeSpecName: "kube-api-access-zmndz") pod "8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" (UID: "8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523"). InnerVolumeSpecName "kube-api-access-zmndz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.500310 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" (UID: "8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.554969 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.577486 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-httpd-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.577518 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmndz\" (UniqueName: \"kubernetes.io/projected/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-kube-api-access-zmndz\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.590621 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" (UID: "8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.623044 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-config" (OuterVolumeSpecName: "config") pod "8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" (UID: "8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.625648 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" (UID: "8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.679931 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.680117 4810 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.680127 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.719703 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c6d89f69d-k9g44"] Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.828829 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d889fdd6d-xxkrj" event={"ID":"8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523","Type":"ContainerDied","Data":"a6c208086449aaef11826ff90109b226e8838be1ee47c86e591f07b3ca7f743a"} Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.828881 4810 scope.go:117] "RemoveContainer" containerID="21d2347bf42c6eb44e7ee7a5de628b64b5a98e56ec1865f17862029f6ca75cc7" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.828991 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d889fdd6d-xxkrj" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.834815 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b","Type":"ContainerStarted","Data":"a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b"} Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.834986 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.835026 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" containerName="cinder-api-log" containerID="cri-o://5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348" gracePeriod=30 Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.835206 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" containerName="cinder-api" containerID="cri-o://a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b" gracePeriod=30 Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.843348 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" event={"ID":"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a","Type":"ContainerStarted","Data":"8e2aa63d354c6a799adc10448416d79da2a260a3c4b244124c9808324a0bad59"} Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.845550 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b753b118-389f-432d-a3a8-366ebb8b440d","Type":"ContainerStarted","Data":"7ea36990f1e8baa8c4bd6d19e5de5041cd332ee987fa1bbf4752442e744c4581"} Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.846846 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c6d89f69d-k9g44" event={"ID":"23454188-6622-4e7c-90f0-2fddf031738e","Type":"ContainerStarted","Data":"1a1661df967ff67a2bed55339e63cfdcac8fe932aad9c6c267f1ac97f8f1f53d"} Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.849598 4810 scope.go:117] "RemoveContainer" containerID="97265a9c186e6460557751232f731c5bc9eac539bb43e0580b1ec02387f960fc" Sep 30 08:22:53 crc kubenswrapper[4810]: E0930 08:22:53.855316 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.867629 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.867608189 podStartE2EDuration="6.867608189s" podCreationTimestamp="2025-09-30 08:22:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:53.850928894 +0000 UTC m=+1197.303128161" watchObservedRunningTime="2025-09-30 08:22:53.867608189 +0000 UTC m=+1197.319807456" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.874510 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.874786 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-687444b9dc-hxqlq" event={"ID":"ff274965-178c-4eff-9841-494634e711f0","Type":"ContainerStarted","Data":"282aacfae9825cd3f08ff249740600690331246ac586dfdacf4e55bda51350bc"} Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.918017 4810 scope.go:117] "RemoveContainer" containerID="d0eeafa4714463a65c8685739ff271f4d79c5316b1d5cc37ad914db6d677ccab" Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.954025 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d889fdd6d-xxkrj"] Sep 30 08:22:53 crc kubenswrapper[4810]: I0930 08:22:53.968404 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d889fdd6d-xxkrj"] Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.002834 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.013533 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.018453 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:22:54 crc kubenswrapper[4810]: E0930 08:22:54.018887 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c545529d-0706-4063-9e3e-75eabc8b9785" containerName="glance-log" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.018904 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c545529d-0706-4063-9e3e-75eabc8b9785" containerName="glance-log" Sep 30 08:22:54 crc kubenswrapper[4810]: E0930 08:22:54.018932 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" containerName="neutron-httpd" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.018939 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" containerName="neutron-httpd" Sep 30 08:22:54 crc kubenswrapper[4810]: E0930 08:22:54.018951 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" containerName="neutron-api" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.018957 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" containerName="neutron-api" Sep 30 08:22:54 crc kubenswrapper[4810]: E0930 08:22:54.018966 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c545529d-0706-4063-9e3e-75eabc8b9785" containerName="glance-httpd" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.018971 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c545529d-0706-4063-9e3e-75eabc8b9785" containerName="glance-httpd" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.019167 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" containerName="neutron-api" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.019187 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c545529d-0706-4063-9e3e-75eabc8b9785" containerName="glance-httpd" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.019204 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" containerName="neutron-httpd" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.019213 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c545529d-0706-4063-9e3e-75eabc8b9785" containerName="glance-log" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.020289 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.022847 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.022969 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.025231 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.187969 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.188772 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.188820 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-logs\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.188888 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.189002 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq67h\" (UniqueName: \"kubernetes.io/projected/69265962-2e8c-430d-b982-208424d9000a-kube-api-access-cq67h\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.189061 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.189103 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.189133 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.290500 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.290582 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.290600 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-logs\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.290625 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.290663 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq67h\" (UniqueName: \"kubernetes.io/projected/69265962-2e8c-430d-b982-208424d9000a-kube-api-access-cq67h\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.290689 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.290712 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.290731 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.292487 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.292728 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-logs\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.293001 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.307983 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.310010 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.310596 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.311019 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.316701 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq67h\" (UniqueName: \"kubernetes.io/projected/69265962-2e8c-430d-b982-208424d9000a-kube-api-access-cq67h\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.349591 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.669901 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.898375 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b753b118-389f-432d-a3a8-366ebb8b440d","Type":"ContainerStarted","Data":"a4b335d34bdb1761d899893ee878fe45cb61a29214a0a773daeea14ea0a7867b"} Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.898720 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b753b118-389f-432d-a3a8-366ebb8b440d","Type":"ContainerStarted","Data":"d0dc8d2cd558ebba2195fb6632f367d95d8e2943e796177366b8402974153dd9"} Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.907939 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c6d89f69d-k9g44" event={"ID":"23454188-6622-4e7c-90f0-2fddf031738e","Type":"ContainerStarted","Data":"161f23282cb5a30b99e4d1eb6303d1c3a763bc8bb995aff87bf6ff84b4ac3f64"} Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.907987 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c6d89f69d-k9g44" event={"ID":"23454188-6622-4e7c-90f0-2fddf031738e","Type":"ContainerStarted","Data":"bfb2c3472313056187f1ec6f3c959d077a9558beaa416ee0d36745badb642b7a"} Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.908175 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.908217 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.928546 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"975d2888-e21f-4ff1-88b4-841427c270e5","Type":"ContainerStarted","Data":"1acf8bfd81f02a2f564cd12a8def41f855578d05ce5835c8c43b9852379cb97c"} Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.930975 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-c6d89f69d-k9g44" podStartSLOduration=2.930957775 podStartE2EDuration="2.930957775s" podCreationTimestamp="2025-09-30 08:22:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:54.92835929 +0000 UTC m=+1198.380558557" watchObservedRunningTime="2025-09-30 08:22:54.930957775 +0000 UTC m=+1198.383157042" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.940967 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-687444b9dc-hxqlq" event={"ID":"ff274965-178c-4eff-9841-494634e711f0","Type":"ContainerStarted","Data":"2780e4663be89cb779d170941400bdfed7770d7d1d8c76257065d41130ceebbe"} Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.961176 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bbc02306-07ec-4a85-95b9-15a4b14b256d","Type":"ContainerStarted","Data":"7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846"} Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.961407 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bbc02306-07ec-4a85-95b9-15a4b14b256d" containerName="glance-log" containerID="cri-o://ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd" gracePeriod=30 Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.961522 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bbc02306-07ec-4a85-95b9-15a4b14b256d" containerName="glance-httpd" containerID="cri-o://7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846" gracePeriod=30 Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.970066 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.374020402 podStartE2EDuration="7.970046501s" podCreationTimestamp="2025-09-30 08:22:47 +0000 UTC" firstStartedPulling="2025-09-30 08:22:48.901768066 +0000 UTC m=+1192.353967333" lastFinishedPulling="2025-09-30 08:22:49.497794165 +0000 UTC m=+1192.949993432" observedRunningTime="2025-09-30 08:22:54.955380895 +0000 UTC m=+1198.407580162" watchObservedRunningTime="2025-09-30 08:22:54.970046501 +0000 UTC m=+1198.422245768" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.980389 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-687444b9dc-hxqlq" podStartSLOduration=5.455381419 podStartE2EDuration="9.980366221s" podCreationTimestamp="2025-09-30 08:22:45 +0000 UTC" firstStartedPulling="2025-09-30 08:22:48.32475234 +0000 UTC m=+1191.776951607" lastFinishedPulling="2025-09-30 08:22:52.849737142 +0000 UTC m=+1196.301936409" observedRunningTime="2025-09-30 08:22:54.975735167 +0000 UTC m=+1198.427934434" watchObservedRunningTime="2025-09-30 08:22:54.980366221 +0000 UTC m=+1198.432565488" Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.986750 4810 generic.go:334] "Generic (PLEG): container finished" podID="79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" containerID="5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348" exitCode=143 Sep 30 08:22:54 crc kubenswrapper[4810]: I0930 08:22:54.986839 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b","Type":"ContainerDied","Data":"5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348"} Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:54.998204 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" event={"ID":"d2c5a1aa-c611-4702-a0bb-27693ebb7c7a","Type":"ContainerStarted","Data":"f338a190d37f5dac991092cb56bb06f14307817f487b519a1ad1f60ede637995"} Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.056819 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=16.056793762 podStartE2EDuration="16.056793762s" podCreationTimestamp="2025-09-30 08:22:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:55.00509912 +0000 UTC m=+1198.457298387" watchObservedRunningTime="2025-09-30 08:22:55.056793762 +0000 UTC m=+1198.508993049" Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.113413 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5dff5594b-jj7gs" podStartSLOduration=5.616525451 podStartE2EDuration="10.113382556s" podCreationTimestamp="2025-09-30 08:22:45 +0000 UTC" firstStartedPulling="2025-09-30 08:22:48.390195151 +0000 UTC m=+1191.842394418" lastFinishedPulling="2025-09-30 08:22:52.887052256 +0000 UTC m=+1196.339251523" observedRunningTime="2025-09-30 08:22:55.033778153 +0000 UTC m=+1198.485977420" watchObservedRunningTime="2025-09-30 08:22:55.113382556 +0000 UTC m=+1198.565581843" Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.326656 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523" path="/var/lib/kubelet/pods/8a8bc0b6-4ebc-4a9f-8ae2-7e6ea5c4e523/volumes" Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.328148 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c545529d-0706-4063-9e3e-75eabc8b9785" path="/var/lib/kubelet/pods/c545529d-0706-4063-9e3e-75eabc8b9785/volumes" Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.334877 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:22:55 crc kubenswrapper[4810]: W0930 08:22:55.349348 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69265962_2e8c_430d_b982_208424d9000a.slice/crio-64bbe17298c76de77325ccf3362517570f2acc52476c8b7170f764c5724d1626 WatchSource:0}: Error finding container 64bbe17298c76de77325ccf3362517570f2acc52476c8b7170f764c5724d1626: Status 404 returned error can't find the container with id 64bbe17298c76de77325ccf3362517570f2acc52476c8b7170f764c5724d1626 Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.733534 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-54b65d888b-2hkpr" Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.859692 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.991033 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"bbc02306-07ec-4a85-95b9-15a4b14b256d\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.991107 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-logs\") pod \"bbc02306-07ec-4a85-95b9-15a4b14b256d\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.991135 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-scripts\") pod \"bbc02306-07ec-4a85-95b9-15a4b14b256d\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.991157 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-config-data\") pod \"bbc02306-07ec-4a85-95b9-15a4b14b256d\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.991237 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-combined-ca-bundle\") pod \"bbc02306-07ec-4a85-95b9-15a4b14b256d\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.991315 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbpxx\" (UniqueName: \"kubernetes.io/projected/bbc02306-07ec-4a85-95b9-15a4b14b256d-kube-api-access-rbpxx\") pod \"bbc02306-07ec-4a85-95b9-15a4b14b256d\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.991392 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-httpd-run\") pod \"bbc02306-07ec-4a85-95b9-15a4b14b256d\" (UID: \"bbc02306-07ec-4a85-95b9-15a4b14b256d\") " Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.992121 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bbc02306-07ec-4a85-95b9-15a4b14b256d" (UID: "bbc02306-07ec-4a85-95b9-15a4b14b256d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.993293 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-logs" (OuterVolumeSpecName: "logs") pod "bbc02306-07ec-4a85-95b9-15a4b14b256d" (UID: "bbc02306-07ec-4a85-95b9-15a4b14b256d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:22:55 crc kubenswrapper[4810]: I0930 08:22:55.999576 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "bbc02306-07ec-4a85-95b9-15a4b14b256d" (UID: "bbc02306-07ec-4a85-95b9-15a4b14b256d"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.006382 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-scripts" (OuterVolumeSpecName: "scripts") pod "bbc02306-07ec-4a85-95b9-15a4b14b256d" (UID: "bbc02306-07ec-4a85-95b9-15a4b14b256d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.020566 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbc02306-07ec-4a85-95b9-15a4b14b256d-kube-api-access-rbpxx" (OuterVolumeSpecName: "kube-api-access-rbpxx") pod "bbc02306-07ec-4a85-95b9-15a4b14b256d" (UID: "bbc02306-07ec-4a85-95b9-15a4b14b256d"). InnerVolumeSpecName "kube-api-access-rbpxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.037677 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbc02306-07ec-4a85-95b9-15a4b14b256d" (UID: "bbc02306-07ec-4a85-95b9-15a4b14b256d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.050745 4810 generic.go:334] "Generic (PLEG): container finished" podID="bbc02306-07ec-4a85-95b9-15a4b14b256d" containerID="7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846" exitCode=0 Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.050782 4810 generic.go:334] "Generic (PLEG): container finished" podID="bbc02306-07ec-4a85-95b9-15a4b14b256d" containerID="ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd" exitCode=143 Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.050826 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bbc02306-07ec-4a85-95b9-15a4b14b256d","Type":"ContainerDied","Data":"7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846"} Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.050854 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bbc02306-07ec-4a85-95b9-15a4b14b256d","Type":"ContainerDied","Data":"ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd"} Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.050866 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bbc02306-07ec-4a85-95b9-15a4b14b256d","Type":"ContainerDied","Data":"6d65c47b8d1307d0ae847fa6237a67377342a1f317c6c57e14b086b0ff86c91a"} Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.050882 4810 scope.go:117] "RemoveContainer" containerID="7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.050999 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.077404 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b753b118-389f-432d-a3a8-366ebb8b440d","Type":"ContainerStarted","Data":"e4178f188246c0cd2bc4f1b754e4d5c857a4a13c3ff2099826b5c21b847de1ea"} Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.085860 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69265962-2e8c-430d-b982-208424d9000a","Type":"ContainerStarted","Data":"64bbe17298c76de77325ccf3362517570f2acc52476c8b7170f764c5724d1626"} Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.094576 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.094809 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbpxx\" (UniqueName: \"kubernetes.io/projected/bbc02306-07ec-4a85-95b9-15a4b14b256d-kube-api-access-rbpxx\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.094879 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.094990 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.095095 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbc02306-07ec-4a85-95b9-15a4b14b256d-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.095155 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.113239 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-config-data" (OuterVolumeSpecName: "config-data") pod "bbc02306-07ec-4a85-95b9-15a4b14b256d" (UID: "bbc02306-07ec-4a85-95b9-15a4b14b256d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.123115 4810 scope.go:117] "RemoveContainer" containerID="ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.137537 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.153456 4810 scope.go:117] "RemoveContainer" containerID="7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846" Sep 30 08:22:56 crc kubenswrapper[4810]: E0930 08:22:56.158379 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846\": container with ID starting with 7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846 not found: ID does not exist" containerID="7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.158421 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846"} err="failed to get container status \"7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846\": rpc error: code = NotFound desc = could not find container \"7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846\": container with ID starting with 7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846 not found: ID does not exist" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.158448 4810 scope.go:117] "RemoveContainer" containerID="ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd" Sep 30 08:22:56 crc kubenswrapper[4810]: E0930 08:22:56.164042 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd\": container with ID starting with ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd not found: ID does not exist" containerID="ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.164091 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd"} err="failed to get container status \"ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd\": rpc error: code = NotFound desc = could not find container \"ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd\": container with ID starting with ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd not found: ID does not exist" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.164125 4810 scope.go:117] "RemoveContainer" containerID="7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.164558 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846"} err="failed to get container status \"7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846\": rpc error: code = NotFound desc = could not find container \"7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846\": container with ID starting with 7076399234dbec1a5cb4eb006e44358135f1d1543b9497153347bc4d532e0846 not found: ID does not exist" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.164578 4810 scope.go:117] "RemoveContainer" containerID="ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.167060 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd"} err="failed to get container status \"ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd\": rpc error: code = NotFound desc = could not find container \"ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd\": container with ID starting with ac9ec2a6c578e08d6815e406eab7f1d4707716bfcf957b3cf141b89f3fcebedd not found: ID does not exist" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.196624 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.196642 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc02306-07ec-4a85-95b9-15a4b14b256d-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.421337 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.435159 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.445069 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:22:56 crc kubenswrapper[4810]: E0930 08:22:56.445535 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbc02306-07ec-4a85-95b9-15a4b14b256d" containerName="glance-log" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.445548 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbc02306-07ec-4a85-95b9-15a4b14b256d" containerName="glance-log" Sep 30 08:22:56 crc kubenswrapper[4810]: E0930 08:22:56.445585 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbc02306-07ec-4a85-95b9-15a4b14b256d" containerName="glance-httpd" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.445592 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbc02306-07ec-4a85-95b9-15a4b14b256d" containerName="glance-httpd" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.445770 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbc02306-07ec-4a85-95b9-15a4b14b256d" containerName="glance-httpd" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.445799 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbc02306-07ec-4a85-95b9-15a4b14b256d" containerName="glance-log" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.446913 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.451309 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.451474 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.462939 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.606329 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-logs\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.606405 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn9vd\" (UniqueName: \"kubernetes.io/projected/852f98b7-5c49-4876-9530-1bacda996d61-kube-api-access-fn9vd\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.606444 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-config-data\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.606465 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-scripts\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.606533 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.606681 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.606819 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.607115 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.709904 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.710155 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.725446 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-logs\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.725545 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn9vd\" (UniqueName: \"kubernetes.io/projected/852f98b7-5c49-4876-9530-1bacda996d61-kube-api-access-fn9vd\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.725585 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-config-data\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.725600 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-scripts\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.726168 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.730421 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.730512 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.732108 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-config-data\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.726513 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-logs\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.733065 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-scripts\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.733388 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.740706 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.762055 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.765987 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn9vd\" (UniqueName: \"kubernetes.io/projected/852f98b7-5c49-4876-9530-1bacda996d61-kube-api-access-fn9vd\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.781842 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " pod="openstack/glance-default-external-api-0" Sep 30 08:22:56 crc kubenswrapper[4810]: I0930 08:22:56.814742 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 08:22:57 crc kubenswrapper[4810]: I0930 08:22:57.125870 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69265962-2e8c-430d-b982-208424d9000a","Type":"ContainerStarted","Data":"d0134ce461d7efd3f22a8c4b2cd16c9053d8b85904f7f219d708c6467c6cd8ab"} Sep 30 08:22:57 crc kubenswrapper[4810]: I0930 08:22:57.126222 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69265962-2e8c-430d-b982-208424d9000a","Type":"ContainerStarted","Data":"c8f8bd13f67c179c83c43fbc5a354363acb5d744b1cc57da1a3f9aa96adc0451"} Sep 30 08:22:57 crc kubenswrapper[4810]: I0930 08:22:57.161839 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.161819478 podStartE2EDuration="4.161819478s" podCreationTimestamp="2025-09-30 08:22:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:22:57.158206923 +0000 UTC m=+1200.610406190" watchObservedRunningTime="2025-09-30 08:22:57.161819478 +0000 UTC m=+1200.614018745" Sep 30 08:22:57 crc kubenswrapper[4810]: I0930 08:22:57.344880 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbc02306-07ec-4a85-95b9-15a4b14b256d" path="/var/lib/kubelet/pods/bbc02306-07ec-4a85-95b9-15a4b14b256d/volumes" Sep 30 08:22:57 crc kubenswrapper[4810]: I0930 08:22:57.472818 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:22:57 crc kubenswrapper[4810]: I0930 08:22:57.965079 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 30 08:22:57 crc kubenswrapper[4810]: I0930 08:22:57.989575 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:22:58 crc kubenswrapper[4810]: I0930 08:22:58.044430 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675956976c-gd7bn"] Sep 30 08:22:58 crc kubenswrapper[4810]: I0930 08:22:58.044721 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-675956976c-gd7bn" podUID="06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" containerName="dnsmasq-dns" containerID="cri-o://2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778" gracePeriod=10 Sep 30 08:22:58 crc kubenswrapper[4810]: I0930 08:22:58.143622 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b753b118-389f-432d-a3a8-366ebb8b440d","Type":"ContainerStarted","Data":"8c33afeca8d86afbe1d3c6f1b99b89e9f56ac12d2cabde574a7cbf37303b540a"} Sep 30 08:22:58 crc kubenswrapper[4810]: I0930 08:22:58.144035 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 08:22:58 crc kubenswrapper[4810]: I0930 08:22:58.152337 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"852f98b7-5c49-4876-9530-1bacda996d61","Type":"ContainerStarted","Data":"d843917bbb0111c02f83752026cec010eefe565c5adde5ffb169c8dec33046f0"} Sep 30 08:22:58 crc kubenswrapper[4810]: I0930 08:22:58.176188 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.312188736 podStartE2EDuration="8.176166491s" podCreationTimestamp="2025-09-30 08:22:50 +0000 UTC" firstStartedPulling="2025-09-30 08:22:53.58340122 +0000 UTC m=+1197.035600487" lastFinishedPulling="2025-09-30 08:22:57.447378985 +0000 UTC m=+1200.899578242" observedRunningTime="2025-09-30 08:22:58.163226405 +0000 UTC m=+1201.615425672" watchObservedRunningTime="2025-09-30 08:22:58.176166491 +0000 UTC m=+1201.628365758" Sep 30 08:22:58 crc kubenswrapper[4810]: I0930 08:22:58.230758 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 30 08:22:58 crc kubenswrapper[4810]: I0930 08:22:58.314335 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 08:22:58 crc kubenswrapper[4810]: I0930 08:22:58.804565 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:58 crc kubenswrapper[4810]: I0930 08:22:58.861397 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.001678 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-config\") pod \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.002422 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-swift-storage-0\") pod \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.002462 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-nb\") pod \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.002503 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-svc\") pod \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.002653 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc28g\" (UniqueName: \"kubernetes.io/projected/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-kube-api-access-nc28g\") pod \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.002770 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-sb\") pod \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.014505 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-kube-api-access-nc28g" (OuterVolumeSpecName: "kube-api-access-nc28g") pod "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" (UID: "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b"). InnerVolumeSpecName "kube-api-access-nc28g". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.106774 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" (UID: "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.107569 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-nb\") pod \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\" (UID: \"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b\") " Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.107966 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc28g\" (UniqueName: \"kubernetes.io/projected/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-kube-api-access-nc28g\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:59 crc kubenswrapper[4810]: W0930 08:22:59.108029 4810 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b/volumes/kubernetes.io~configmap/ovsdbserver-nb Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.108037 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" (UID: "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.115894 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" (UID: "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.116337 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" (UID: "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.125704 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-config" (OuterVolumeSpecName: "config") pod "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" (UID: "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.138128 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.143830 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" (UID: "06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.198686 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"852f98b7-5c49-4876-9530-1bacda996d61","Type":"ContainerStarted","Data":"6c5fdae1523b62cd2c8f8055eb375f1729321df4f5a89cbd08c1074f4cac39f4"} Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.209528 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.209552 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.209562 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.209574 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.209583 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.224696 4810 generic.go:334] "Generic (PLEG): container finished" podID="06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" containerID="2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778" exitCode=0 Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.225683 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675956976c-gd7bn" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.227822 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675956976c-gd7bn" event={"ID":"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b","Type":"ContainerDied","Data":"2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778"} Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.227857 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675956976c-gd7bn" event={"ID":"06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b","Type":"ContainerDied","Data":"04e24be23db9db89fa11356285883a2c3b6b25aefd35d153252aa06995827029"} Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.227875 4810 scope.go:117] "RemoveContainer" containerID="2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.228092 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="975d2888-e21f-4ff1-88b4-841427c270e5" containerName="cinder-scheduler" containerID="cri-o://99541a51d283a02a348a3ddbbbe4d6b66ebcdec80efa338931ab2ff8c1db9866" gracePeriod=30 Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.228184 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="975d2888-e21f-4ff1-88b4-841427c270e5" containerName="probe" containerID="cri-o://1acf8bfd81f02a2f564cd12a8def41f855578d05ce5835c8c43b9852379cb97c" gracePeriod=30 Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.295216 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675956976c-gd7bn"] Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.329518 4810 scope.go:117] "RemoveContainer" containerID="a220086bf118a538d86466e7606c6504bcba26f94ee3793ecd9f4407fc73a027" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.350001 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675956976c-gd7bn"] Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.428534 4810 scope.go:117] "RemoveContainer" containerID="2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778" Sep 30 08:22:59 crc kubenswrapper[4810]: E0930 08:22:59.429324 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778\": container with ID starting with 2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778 not found: ID does not exist" containerID="2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.429366 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778"} err="failed to get container status \"2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778\": rpc error: code = NotFound desc = could not find container \"2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778\": container with ID starting with 2ea4096c92032c3dc4bc7338155547213abe109e328cea30d99d95986c89d778 not found: ID does not exist" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.429400 4810 scope.go:117] "RemoveContainer" containerID="a220086bf118a538d86466e7606c6504bcba26f94ee3793ecd9f4407fc73a027" Sep 30 08:22:59 crc kubenswrapper[4810]: E0930 08:22:59.431489 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a220086bf118a538d86466e7606c6504bcba26f94ee3793ecd9f4407fc73a027\": container with ID starting with a220086bf118a538d86466e7606c6504bcba26f94ee3793ecd9f4407fc73a027 not found: ID does not exist" containerID="a220086bf118a538d86466e7606c6504bcba26f94ee3793ecd9f4407fc73a027" Sep 30 08:22:59 crc kubenswrapper[4810]: I0930 08:22:59.431524 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a220086bf118a538d86466e7606c6504bcba26f94ee3793ecd9f4407fc73a027"} err="failed to get container status \"a220086bf118a538d86466e7606c6504bcba26f94ee3793ecd9f4407fc73a027\": rpc error: code = NotFound desc = could not find container \"a220086bf118a538d86466e7606c6504bcba26f94ee3793ecd9f4407fc73a027\": container with ID starting with a220086bf118a538d86466e7606c6504bcba26f94ee3793ecd9f4407fc73a027 not found: ID does not exist" Sep 30 08:23:00 crc kubenswrapper[4810]: I0930 08:23:00.238490 4810 generic.go:334] "Generic (PLEG): container finished" podID="975d2888-e21f-4ff1-88b4-841427c270e5" containerID="1acf8bfd81f02a2f564cd12a8def41f855578d05ce5835c8c43b9852379cb97c" exitCode=0 Sep 30 08:23:00 crc kubenswrapper[4810]: I0930 08:23:00.238556 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"975d2888-e21f-4ff1-88b4-841427c270e5","Type":"ContainerDied","Data":"1acf8bfd81f02a2f564cd12a8def41f855578d05ce5835c8c43b9852379cb97c"} Sep 30 08:23:00 crc kubenswrapper[4810]: I0930 08:23:00.242509 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"852f98b7-5c49-4876-9530-1bacda996d61","Type":"ContainerStarted","Data":"ce6ef8f3ade002003120fee17fee20ef6302aebca084e8c99dcb4fe6d1102e3b"} Sep 30 08:23:00 crc kubenswrapper[4810]: I0930 08:23:00.265629 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.265611604 podStartE2EDuration="4.265611604s" podCreationTimestamp="2025-09-30 08:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:23:00.264125301 +0000 UTC m=+1203.716324568" watchObservedRunningTime="2025-09-30 08:23:00.265611604 +0000 UTC m=+1203.717810871" Sep 30 08:23:00 crc kubenswrapper[4810]: I0930 08:23:00.468957 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-f95cc94f6-tz5ks" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.252594 4810 generic.go:334] "Generic (PLEG): container finished" podID="975d2888-e21f-4ff1-88b4-841427c270e5" containerID="99541a51d283a02a348a3ddbbbe4d6b66ebcdec80efa338931ab2ff8c1db9866" exitCode=0 Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.252677 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"975d2888-e21f-4ff1-88b4-841427c270e5","Type":"ContainerDied","Data":"99541a51d283a02a348a3ddbbbe4d6b66ebcdec80efa338931ab2ff8c1db9866"} Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.253319 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"975d2888-e21f-4ff1-88b4-841427c270e5","Type":"ContainerDied","Data":"842fba3d2eb546de17ec1de7455e43a4d477fc885a73207d870531d8c4c1f7ea"} Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.253336 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="842fba3d2eb546de17ec1de7455e43a4d477fc885a73207d870531d8c4c1f7ea" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.308667 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.320079 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" path="/var/lib/kubelet/pods/06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b/volumes" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.353467 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksnxn\" (UniqueName: \"kubernetes.io/projected/975d2888-e21f-4ff1-88b4-841427c270e5-kube-api-access-ksnxn\") pod \"975d2888-e21f-4ff1-88b4-841427c270e5\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.353579 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/975d2888-e21f-4ff1-88b4-841427c270e5-etc-machine-id\") pod \"975d2888-e21f-4ff1-88b4-841427c270e5\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.353623 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data-custom\") pod \"975d2888-e21f-4ff1-88b4-841427c270e5\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.353665 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-combined-ca-bundle\") pod \"975d2888-e21f-4ff1-88b4-841427c270e5\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.353685 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data\") pod \"975d2888-e21f-4ff1-88b4-841427c270e5\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.353737 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-scripts\") pod \"975d2888-e21f-4ff1-88b4-841427c270e5\" (UID: \"975d2888-e21f-4ff1-88b4-841427c270e5\") " Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.355731 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/975d2888-e21f-4ff1-88b4-841427c270e5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "975d2888-e21f-4ff1-88b4-841427c270e5" (UID: "975d2888-e21f-4ff1-88b4-841427c270e5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.371317 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "975d2888-e21f-4ff1-88b4-841427c270e5" (UID: "975d2888-e21f-4ff1-88b4-841427c270e5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.372845 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/975d2888-e21f-4ff1-88b4-841427c270e5-kube-api-access-ksnxn" (OuterVolumeSpecName: "kube-api-access-ksnxn") pod "975d2888-e21f-4ff1-88b4-841427c270e5" (UID: "975d2888-e21f-4ff1-88b4-841427c270e5"). InnerVolumeSpecName "kube-api-access-ksnxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.374690 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-scripts" (OuterVolumeSpecName: "scripts") pod "975d2888-e21f-4ff1-88b4-841427c270e5" (UID: "975d2888-e21f-4ff1-88b4-841427c270e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.428121 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.440570 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "975d2888-e21f-4ff1-88b4-841427c270e5" (UID: "975d2888-e21f-4ff1-88b4-841427c270e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.468200 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksnxn\" (UniqueName: \"kubernetes.io/projected/975d2888-e21f-4ff1-88b4-841427c270e5-kube-api-access-ksnxn\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.468246 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/975d2888-e21f-4ff1-88b4-841427c270e5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.468259 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.468284 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.468294 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.509229 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data" (OuterVolumeSpecName: "config-data") pod "975d2888-e21f-4ff1-88b4-841427c270e5" (UID: "975d2888-e21f-4ff1-88b4-841427c270e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:01 crc kubenswrapper[4810]: I0930 08:23:01.569843 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975d2888-e21f-4ff1-88b4-841427c270e5-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.260794 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.292091 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.299027 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.314534 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 08:23:02 crc kubenswrapper[4810]: E0930 08:23:02.314913 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975d2888-e21f-4ff1-88b4-841427c270e5" containerName="cinder-scheduler" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.314932 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="975d2888-e21f-4ff1-88b4-841427c270e5" containerName="cinder-scheduler" Sep 30 08:23:02 crc kubenswrapper[4810]: E0930 08:23:02.314954 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975d2888-e21f-4ff1-88b4-841427c270e5" containerName="probe" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.314962 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="975d2888-e21f-4ff1-88b4-841427c270e5" containerName="probe" Sep 30 08:23:02 crc kubenswrapper[4810]: E0930 08:23:02.314972 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" containerName="dnsmasq-dns" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.314978 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" containerName="dnsmasq-dns" Sep 30 08:23:02 crc kubenswrapper[4810]: E0930 08:23:02.314987 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" containerName="init" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.314992 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" containerName="init" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.315173 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b" containerName="dnsmasq-dns" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.315186 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="975d2888-e21f-4ff1-88b4-841427c270e5" containerName="probe" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.315208 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="975d2888-e21f-4ff1-88b4-841427c270e5" containerName="cinder-scheduler" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.316463 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.318390 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.332732 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.386934 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e10f507-33b5-4477-882f-6c0b3271034d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.387249 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vxr7\" (UniqueName: \"kubernetes.io/projected/2e10f507-33b5-4477-882f-6c0b3271034d-kube-api-access-4vxr7\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.387380 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.387663 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.387792 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.387914 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.489863 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.489966 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.490002 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.490041 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.490100 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e10f507-33b5-4477-882f-6c0b3271034d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.490119 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vxr7\" (UniqueName: \"kubernetes.io/projected/2e10f507-33b5-4477-882f-6c0b3271034d-kube-api-access-4vxr7\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.491491 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e10f507-33b5-4477-882f-6c0b3271034d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.495010 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.496159 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.500012 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.503743 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e10f507-33b5-4477-882f-6c0b3271034d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.517693 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vxr7\" (UniqueName: \"kubernetes.io/projected/2e10f507-33b5-4477-882f-6c0b3271034d-kube-api-access-4vxr7\") pod \"cinder-scheduler-0\" (UID: \"2e10f507-33b5-4477-882f-6c0b3271034d\") " pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.566844 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.568097 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.570298 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.570451 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-26bpc" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.570509 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.584864 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.638551 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.693117 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-openstack-config-secret\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.693665 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tltr\" (UniqueName: \"kubernetes.io/projected/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-kube-api-access-8tltr\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.693742 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.693886 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-openstack-config\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.779215 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-754cbc7dbb-wbdt6" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.163:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.163:8443: connect: connection refused" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.779348 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.801845 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tltr\" (UniqueName: \"kubernetes.io/projected/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-kube-api-access-8tltr\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.801966 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.802143 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-openstack-config\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.802183 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-openstack-config-secret\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.808301 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-openstack-config-secret\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.809152 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-openstack-config\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.812959 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.825903 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tltr\" (UniqueName: \"kubernetes.io/projected/f0aa786d-3e79-423b-82d1-5e9c70ca7c72-kube-api-access-8tltr\") pod \"openstackclient\" (UID: \"f0aa786d-3e79-423b-82d1-5e9c70ca7c72\") " pod="openstack/openstackclient" Sep 30 08:23:02 crc kubenswrapper[4810]: I0930 08:23:02.918759 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 08:23:03 crc kubenswrapper[4810]: I0930 08:23:03.155075 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 08:23:03 crc kubenswrapper[4810]: I0930 08:23:03.285104 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e10f507-33b5-4477-882f-6c0b3271034d","Type":"ContainerStarted","Data":"1600287ba1953e54f41a7ff14d990093e3484b6c23726e0fee2fc12cc8790c6a"} Sep 30 08:23:03 crc kubenswrapper[4810]: I0930 08:23:03.317191 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="975d2888-e21f-4ff1-88b4-841427c270e5" path="/var/lib/kubelet/pods/975d2888-e21f-4ff1-88b4-841427c270e5/volumes" Sep 30 08:23:03 crc kubenswrapper[4810]: I0930 08:23:03.480085 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 30 08:23:04 crc kubenswrapper[4810]: I0930 08:23:04.309833 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f0aa786d-3e79-423b-82d1-5e9c70ca7c72","Type":"ContainerStarted","Data":"72f8c785087de0ad6477700d0ac9b5f57b10bb721ccb122fcc31569dd19526d0"} Sep 30 08:23:04 crc kubenswrapper[4810]: I0930 08:23:04.315074 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e10f507-33b5-4477-882f-6c0b3271034d","Type":"ContainerStarted","Data":"e04b57cc0ab5092f531914cb8137397832bace5ac80a5410639df879739ddaaf"} Sep 30 08:23:04 crc kubenswrapper[4810]: I0930 08:23:04.671885 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:04 crc kubenswrapper[4810]: I0930 08:23:04.672150 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:04 crc kubenswrapper[4810]: I0930 08:23:04.715369 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:04 crc kubenswrapper[4810]: I0930 08:23:04.743721 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:23:04 crc kubenswrapper[4810]: I0930 08:23:04.750120 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:04 crc kubenswrapper[4810]: I0930 08:23:04.873226 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c6d89f69d-k9g44" Sep 30 08:23:04 crc kubenswrapper[4810]: I0930 08:23:04.953831 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65d8f9c97d-267vl"] Sep 30 08:23:04 crc kubenswrapper[4810]: I0930 08:23:04.954062 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65d8f9c97d-267vl" podUID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerName="barbican-api-log" containerID="cri-o://cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491" gracePeriod=30 Sep 30 08:23:04 crc kubenswrapper[4810]: I0930 08:23:04.954462 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65d8f9c97d-267vl" podUID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerName="barbican-api" containerID="cri-o://fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351" gracePeriod=30 Sep 30 08:23:05 crc kubenswrapper[4810]: I0930 08:23:05.358550 4810 generic.go:334] "Generic (PLEG): container finished" podID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerID="cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491" exitCode=143 Sep 30 08:23:05 crc kubenswrapper[4810]: I0930 08:23:05.358615 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65d8f9c97d-267vl" event={"ID":"5ac68abd-cf5c-4a46-9db1-599d547c3b06","Type":"ContainerDied","Data":"cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491"} Sep 30 08:23:05 crc kubenswrapper[4810]: I0930 08:23:05.376053 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e10f507-33b5-4477-882f-6c0b3271034d","Type":"ContainerStarted","Data":"3c57bdc987d94ff3b41e4265b30894e5d94866820f3ee888088de37f6eeea6b0"} Sep 30 08:23:05 crc kubenswrapper[4810]: I0930 08:23:05.376101 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:05 crc kubenswrapper[4810]: I0930 08:23:05.376733 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:05 crc kubenswrapper[4810]: I0930 08:23:05.410507 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.410481757 podStartE2EDuration="3.410481757s" podCreationTimestamp="2025-09-30 08:23:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:23:05.399738125 +0000 UTC m=+1208.851937392" watchObservedRunningTime="2025-09-30 08:23:05.410481757 +0000 UTC m=+1208.862681044" Sep 30 08:23:06 crc kubenswrapper[4810]: I0930 08:23:06.307109 4810 scope.go:117] "RemoveContainer" containerID="97265a9c186e6460557751232f731c5bc9eac539bb43e0580b1ec02387f960fc" Sep 30 08:23:06 crc kubenswrapper[4810]: E0930 08:23:06.307642 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:23:06 crc kubenswrapper[4810]: I0930 08:23:06.568083 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65d8f9c97d-267vl" podUID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.180:9311/healthcheck\": read tcp 10.217.0.2:45176->10.217.0.180:9311: read: connection reset by peer" Sep 30 08:23:06 crc kubenswrapper[4810]: I0930 08:23:06.568225 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65d8f9c97d-267vl" podUID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.180:9311/healthcheck\": read tcp 10.217.0.2:45192->10.217.0.180:9311: read: connection reset by peer" Sep 30 08:23:06 crc kubenswrapper[4810]: I0930 08:23:06.815575 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 08:23:06 crc kubenswrapper[4810]: I0930 08:23:06.815838 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 08:23:06 crc kubenswrapper[4810]: I0930 08:23:06.859070 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 08:23:06 crc kubenswrapper[4810]: I0930 08:23:06.859422 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.296210 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.354302 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data\") pod \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.354502 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ac68abd-cf5c-4a46-9db1-599d547c3b06-logs\") pod \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.354609 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xczhb\" (UniqueName: \"kubernetes.io/projected/5ac68abd-cf5c-4a46-9db1-599d547c3b06-kube-api-access-xczhb\") pod \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.354627 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-combined-ca-bundle\") pod \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.354652 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data-custom\") pod \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\" (UID: \"5ac68abd-cf5c-4a46-9db1-599d547c3b06\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.356447 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ac68abd-cf5c-4a46-9db1-599d547c3b06-logs" (OuterVolumeSpecName: "logs") pod "5ac68abd-cf5c-4a46-9db1-599d547c3b06" (UID: "5ac68abd-cf5c-4a46-9db1-599d547c3b06"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.361300 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ac68abd-cf5c-4a46-9db1-599d547c3b06-kube-api-access-xczhb" (OuterVolumeSpecName: "kube-api-access-xczhb") pod "5ac68abd-cf5c-4a46-9db1-599d547c3b06" (UID: "5ac68abd-cf5c-4a46-9db1-599d547c3b06"). InnerVolumeSpecName "kube-api-access-xczhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.361588 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5ac68abd-cf5c-4a46-9db1-599d547c3b06" (UID: "5ac68abd-cf5c-4a46-9db1-599d547c3b06"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.384155 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ac68abd-cf5c-4a46-9db1-599d547c3b06" (UID: "5ac68abd-cf5c-4a46-9db1-599d547c3b06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.398832 4810 generic.go:334] "Generic (PLEG): container finished" podID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerID="fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351" exitCode=0 Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.398944 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65d8f9c97d-267vl" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.398900 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65d8f9c97d-267vl" event={"ID":"5ac68abd-cf5c-4a46-9db1-599d547c3b06","Type":"ContainerDied","Data":"fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351"} Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.399065 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65d8f9c97d-267vl" event={"ID":"5ac68abd-cf5c-4a46-9db1-599d547c3b06","Type":"ContainerDied","Data":"730c6d8f1bd112a78dda99f986a7691880de092ca2d81faef317b0f3a0882db4"} Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.399086 4810 scope.go:117] "RemoveContainer" containerID="fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.407692 4810 generic.go:334] "Generic (PLEG): container finished" podID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerID="65c41b5d112da8a316221d00dc0b08a61accf182470442b32f5aa0e475f2698f" exitCode=137 Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.407778 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754cbc7dbb-wbdt6" event={"ID":"c176d78f-1abf-4191-94f8-8ee7be102b18","Type":"ContainerDied","Data":"65c41b5d112da8a316221d00dc0b08a61accf182470442b32f5aa0e475f2698f"} Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.408018 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.408046 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.408589 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.408616 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.416740 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data" (OuterVolumeSpecName: "config-data") pod "5ac68abd-cf5c-4a46-9db1-599d547c3b06" (UID: "5ac68abd-cf5c-4a46-9db1-599d547c3b06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.456632 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xczhb\" (UniqueName: \"kubernetes.io/projected/5ac68abd-cf5c-4a46-9db1-599d547c3b06-kube-api-access-xczhb\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.456668 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.456677 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.456687 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ac68abd-cf5c-4a46-9db1-599d547c3b06-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.456697 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ac68abd-cf5c-4a46-9db1-599d547c3b06-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.475094 4810 scope.go:117] "RemoveContainer" containerID="cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.500253 4810 scope.go:117] "RemoveContainer" containerID="fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351" Sep 30 08:23:07 crc kubenswrapper[4810]: E0930 08:23:07.500849 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351\": container with ID starting with fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351 not found: ID does not exist" containerID="fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.500881 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351"} err="failed to get container status \"fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351\": rpc error: code = NotFound desc = could not find container \"fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351\": container with ID starting with fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351 not found: ID does not exist" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.500902 4810 scope.go:117] "RemoveContainer" containerID="cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491" Sep 30 08:23:07 crc kubenswrapper[4810]: E0930 08:23:07.501967 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491\": container with ID starting with cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491 not found: ID does not exist" containerID="cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.501993 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491"} err="failed to get container status \"cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491\": rpc error: code = NotFound desc = could not find container \"cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491\": container with ID starting with cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491 not found: ID does not exist" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.576578 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.638803 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.672523 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c176d78f-1abf-4191-94f8-8ee7be102b18-logs\") pod \"c176d78f-1abf-4191-94f8-8ee7be102b18\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.672600 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6tx2\" (UniqueName: \"kubernetes.io/projected/c176d78f-1abf-4191-94f8-8ee7be102b18-kube-api-access-q6tx2\") pod \"c176d78f-1abf-4191-94f8-8ee7be102b18\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.672650 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-tls-certs\") pod \"c176d78f-1abf-4191-94f8-8ee7be102b18\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.672692 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-secret-key\") pod \"c176d78f-1abf-4191-94f8-8ee7be102b18\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.672778 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-combined-ca-bundle\") pod \"c176d78f-1abf-4191-94f8-8ee7be102b18\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.672835 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-scripts\") pod \"c176d78f-1abf-4191-94f8-8ee7be102b18\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.672871 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-config-data\") pod \"c176d78f-1abf-4191-94f8-8ee7be102b18\" (UID: \"c176d78f-1abf-4191-94f8-8ee7be102b18\") " Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.678368 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c176d78f-1abf-4191-94f8-8ee7be102b18-logs" (OuterVolumeSpecName: "logs") pod "c176d78f-1abf-4191-94f8-8ee7be102b18" (UID: "c176d78f-1abf-4191-94f8-8ee7be102b18"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.683263 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c176d78f-1abf-4191-94f8-8ee7be102b18-kube-api-access-q6tx2" (OuterVolumeSpecName: "kube-api-access-q6tx2") pod "c176d78f-1abf-4191-94f8-8ee7be102b18" (UID: "c176d78f-1abf-4191-94f8-8ee7be102b18"). InnerVolumeSpecName "kube-api-access-q6tx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.693935 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c176d78f-1abf-4191-94f8-8ee7be102b18" (UID: "c176d78f-1abf-4191-94f8-8ee7be102b18"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.726149 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-config-data" (OuterVolumeSpecName: "config-data") pod "c176d78f-1abf-4191-94f8-8ee7be102b18" (UID: "c176d78f-1abf-4191-94f8-8ee7be102b18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.739431 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "c176d78f-1abf-4191-94f8-8ee7be102b18" (UID: "c176d78f-1abf-4191-94f8-8ee7be102b18"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.750841 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-scripts" (OuterVolumeSpecName: "scripts") pod "c176d78f-1abf-4191-94f8-8ee7be102b18" (UID: "c176d78f-1abf-4191-94f8-8ee7be102b18"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.751000 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c176d78f-1abf-4191-94f8-8ee7be102b18" (UID: "c176d78f-1abf-4191-94f8-8ee7be102b18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.776826 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.776865 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c176d78f-1abf-4191-94f8-8ee7be102b18-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.776878 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6tx2\" (UniqueName: \"kubernetes.io/projected/c176d78f-1abf-4191-94f8-8ee7be102b18-kube-api-access-q6tx2\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.776889 4810 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.776899 4810 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.776909 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c176d78f-1abf-4191-94f8-8ee7be102b18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.776918 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c176d78f-1abf-4191-94f8-8ee7be102b18-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.848240 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65d8f9c97d-267vl"] Sep 30 08:23:07 crc kubenswrapper[4810]: I0930 08:23:07.858102 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-65d8f9c97d-267vl"] Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.419315 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-754cbc7dbb-wbdt6" event={"ID":"c176d78f-1abf-4191-94f8-8ee7be102b18","Type":"ContainerDied","Data":"c4a402a0d631408a6d207e0bd44625463f0d1368dd2485a3e052ac7d06864225"} Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.419347 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-754cbc7dbb-wbdt6" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.419388 4810 scope.go:117] "RemoveContainer" containerID="a85450da47bf890fcf9ae454a6f57b37faf86b24f277f113255feb929be07b4c" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.462821 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-754cbc7dbb-wbdt6"] Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.472290 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-754cbc7dbb-wbdt6"] Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.612541 4810 scope.go:117] "RemoveContainer" containerID="65c41b5d112da8a316221d00dc0b08a61accf182470442b32f5aa0e475f2698f" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.789887 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5cd556568c-9c6km"] Sep 30 08:23:08 crc kubenswrapper[4810]: E0930 08:23:08.790287 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerName="horizon-log" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.790303 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerName="horizon-log" Sep 30 08:23:08 crc kubenswrapper[4810]: E0930 08:23:08.790314 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerName="barbican-api" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.790320 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerName="barbican-api" Sep 30 08:23:08 crc kubenswrapper[4810]: E0930 08:23:08.790332 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerName="horizon" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.790338 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerName="horizon" Sep 30 08:23:08 crc kubenswrapper[4810]: E0930 08:23:08.790372 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerName="barbican-api-log" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.790379 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerName="barbican-api-log" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.790560 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerName="barbican-api" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.790572 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerName="horizon" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.790593 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" containerName="horizon-log" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.790616 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" containerName="barbican-api-log" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.791624 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.795701 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.795898 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.799181 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.814025 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5cd556568c-9c6km"] Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.898781 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-combined-ca-bundle\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.899176 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc1aa89c-399a-4db2-a0b1-df1185b14c48-run-httpd\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.899240 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc1aa89c-399a-4db2-a0b1-df1185b14c48-log-httpd\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.899390 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/dc1aa89c-399a-4db2-a0b1-df1185b14c48-etc-swift\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.899443 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-internal-tls-certs\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.899770 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-public-tls-certs\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.899819 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk58l\" (UniqueName: \"kubernetes.io/projected/dc1aa89c-399a-4db2-a0b1-df1185b14c48-kube-api-access-xk58l\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:08 crc kubenswrapper[4810]: I0930 08:23:08.899881 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-config-data\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.008344 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/dc1aa89c-399a-4db2-a0b1-df1185b14c48-etc-swift\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.008409 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-internal-tls-certs\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.008518 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-public-tls-certs\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.008534 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk58l\" (UniqueName: \"kubernetes.io/projected/dc1aa89c-399a-4db2-a0b1-df1185b14c48-kube-api-access-xk58l\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.008554 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-config-data\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.008582 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-combined-ca-bundle\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.008605 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc1aa89c-399a-4db2-a0b1-df1185b14c48-run-httpd\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.008625 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc1aa89c-399a-4db2-a0b1-df1185b14c48-log-httpd\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.009116 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc1aa89c-399a-4db2-a0b1-df1185b14c48-log-httpd\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.014706 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-public-tls-certs\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.017752 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc1aa89c-399a-4db2-a0b1-df1185b14c48-run-httpd\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.021939 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-combined-ca-bundle\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.024956 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-config-data\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.026697 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc1aa89c-399a-4db2-a0b1-df1185b14c48-internal-tls-certs\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.047646 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/dc1aa89c-399a-4db2-a0b1-df1185b14c48-etc-swift\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.054337 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk58l\" (UniqueName: \"kubernetes.io/projected/dc1aa89c-399a-4db2-a0b1-df1185b14c48-kube-api-access-xk58l\") pod \"swift-proxy-5cd556568c-9c6km\" (UID: \"dc1aa89c-399a-4db2-a0b1-df1185b14c48\") " pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.114707 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.326165 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ac68abd-cf5c-4a46-9db1-599d547c3b06" path="/var/lib/kubelet/pods/5ac68abd-cf5c-4a46-9db1-599d547c3b06/volumes" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.327208 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c176d78f-1abf-4191-94f8-8ee7be102b18" path="/var/lib/kubelet/pods/c176d78f-1abf-4191-94f8-8ee7be102b18/volumes" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.793526 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5cd556568c-9c6km"] Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.827508 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.827609 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.847028 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.869021 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.869157 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 08:23:09 crc kubenswrapper[4810]: I0930 08:23:09.871043 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.182372 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.182789 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="proxy-httpd" containerID="cri-o://8c33afeca8d86afbe1d3c6f1b99b89e9f56ac12d2cabde574a7cbf37303b540a" gracePeriod=30 Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.182872 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="sg-core" containerID="cri-o://e4178f188246c0cd2bc4f1b754e4d5c857a4a13c3ff2099826b5c21b847de1ea" gracePeriod=30 Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.182980 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="ceilometer-central-agent" containerID="cri-o://d0dc8d2cd558ebba2195fb6632f367d95d8e2943e796177366b8402974153dd9" gracePeriod=30 Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.183042 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="ceilometer-notification-agent" containerID="cri-o://a4b335d34bdb1761d899893ee878fe45cb61a29214a0a773daeea14ea0a7867b" gracePeriod=30 Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.195726 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.184:3000/\": EOF" Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.459656 4810 generic.go:334] "Generic (PLEG): container finished" podID="b753b118-389f-432d-a3a8-366ebb8b440d" containerID="8c33afeca8d86afbe1d3c6f1b99b89e9f56ac12d2cabde574a7cbf37303b540a" exitCode=0 Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.459695 4810 generic.go:334] "Generic (PLEG): container finished" podID="b753b118-389f-432d-a3a8-366ebb8b440d" containerID="e4178f188246c0cd2bc4f1b754e4d5c857a4a13c3ff2099826b5c21b847de1ea" exitCode=2 Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.459750 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b753b118-389f-432d-a3a8-366ebb8b440d","Type":"ContainerDied","Data":"8c33afeca8d86afbe1d3c6f1b99b89e9f56ac12d2cabde574a7cbf37303b540a"} Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.459780 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b753b118-389f-432d-a3a8-366ebb8b440d","Type":"ContainerDied","Data":"e4178f188246c0cd2bc4f1b754e4d5c857a4a13c3ff2099826b5c21b847de1ea"} Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.461835 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5cd556568c-9c6km" event={"ID":"dc1aa89c-399a-4db2-a0b1-df1185b14c48","Type":"ContainerStarted","Data":"9a268bf8d7046457a2ab59e0b20f25c9b4233794d47e9f21edeff1a91a09ae79"} Sep 30 08:23:10 crc kubenswrapper[4810]: I0930 08:23:10.461898 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5cd556568c-9c6km" event={"ID":"dc1aa89c-399a-4db2-a0b1-df1185b14c48","Type":"ContainerStarted","Data":"191b2e6f25a2f326de94101a880b349686e6154fc50d5a7d4e78b679fcb628f1"} Sep 30 08:23:11 crc kubenswrapper[4810]: I0930 08:23:11.475462 4810 generic.go:334] "Generic (PLEG): container finished" podID="b753b118-389f-432d-a3a8-366ebb8b440d" containerID="a4b335d34bdb1761d899893ee878fe45cb61a29214a0a773daeea14ea0a7867b" exitCode=0 Sep 30 08:23:11 crc kubenswrapper[4810]: I0930 08:23:11.475772 4810 generic.go:334] "Generic (PLEG): container finished" podID="b753b118-389f-432d-a3a8-366ebb8b440d" containerID="d0dc8d2cd558ebba2195fb6632f367d95d8e2943e796177366b8402974153dd9" exitCode=0 Sep 30 08:23:11 crc kubenswrapper[4810]: I0930 08:23:11.475825 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b753b118-389f-432d-a3a8-366ebb8b440d","Type":"ContainerDied","Data":"a4b335d34bdb1761d899893ee878fe45cb61a29214a0a773daeea14ea0a7867b"} Sep 30 08:23:11 crc kubenswrapper[4810]: I0930 08:23:11.475861 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b753b118-389f-432d-a3a8-366ebb8b440d","Type":"ContainerDied","Data":"d0dc8d2cd558ebba2195fb6632f367d95d8e2943e796177366b8402974153dd9"} Sep 30 08:23:11 crc kubenswrapper[4810]: I0930 08:23:11.479245 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5cd556568c-9c6km" event={"ID":"dc1aa89c-399a-4db2-a0b1-df1185b14c48","Type":"ContainerStarted","Data":"687ce5dd6acac5d2747f16d1c0f3bad4f45334ac17e975ac0d706560212c25e6"} Sep 30 08:23:11 crc kubenswrapper[4810]: I0930 08:23:11.479666 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:11 crc kubenswrapper[4810]: I0930 08:23:11.479699 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:11 crc kubenswrapper[4810]: I0930 08:23:11.506799 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5cd556568c-9c6km" podStartSLOduration=3.506777627 podStartE2EDuration="3.506777627s" podCreationTimestamp="2025-09-30 08:23:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:23:11.505543251 +0000 UTC m=+1214.957742518" watchObservedRunningTime="2025-09-30 08:23:11.506777627 +0000 UTC m=+1214.958976894" Sep 30 08:23:12 crc kubenswrapper[4810]: I0930 08:23:12.906993 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 30 08:23:13 crc kubenswrapper[4810]: I0930 08:23:13.478121 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Sep 30 08:23:13 crc kubenswrapper[4810]: I0930 08:23:13.478465 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:23:13 crc kubenswrapper[4810]: I0930 08:23:13.479279 4810 scope.go:117] "RemoveContainer" containerID="97265a9c186e6460557751232f731c5bc9eac539bb43e0580b1ec02387f960fc" Sep 30 08:23:15 crc kubenswrapper[4810]: I0930 08:23:15.327908 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:23:15 crc kubenswrapper[4810]: I0930 08:23:15.328475 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="69265962-2e8c-430d-b982-208424d9000a" containerName="glance-log" containerID="cri-o://c8f8bd13f67c179c83c43fbc5a354363acb5d744b1cc57da1a3f9aa96adc0451" gracePeriod=30 Sep 30 08:23:15 crc kubenswrapper[4810]: I0930 08:23:15.328854 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="69265962-2e8c-430d-b982-208424d9000a" containerName="glance-httpd" containerID="cri-o://d0134ce461d7efd3f22a8c4b2cd16c9053d8b85904f7f219d708c6467c6cd8ab" gracePeriod=30 Sep 30 08:23:15 crc kubenswrapper[4810]: I0930 08:23:15.524309 4810 generic.go:334] "Generic (PLEG): container finished" podID="69265962-2e8c-430d-b982-208424d9000a" containerID="c8f8bd13f67c179c83c43fbc5a354363acb5d744b1cc57da1a3f9aa96adc0451" exitCode=143 Sep 30 08:23:15 crc kubenswrapper[4810]: I0930 08:23:15.524357 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69265962-2e8c-430d-b982-208424d9000a","Type":"ContainerDied","Data":"c8f8bd13f67c179c83c43fbc5a354363acb5d744b1cc57da1a3f9aa96adc0451"} Sep 30 08:23:15 crc kubenswrapper[4810]: I0930 08:23:15.912017 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:23:15 crc kubenswrapper[4810]: I0930 08:23:15.912097 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:23:16 crc kubenswrapper[4810]: E0930 08:23:16.372250 4810 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/bcb149087e793ddcf5e24488bb613b1c5ec8aada96bc45859df5b13394ab991d/diff" to get inode usage: stat /var/lib/containers/storage/overlay/bcb149087e793ddcf5e24488bb613b1c5ec8aada96bc45859df5b13394ab991d/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_horizon-754cbc7dbb-wbdt6_c176d78f-1abf-4191-94f8-8ee7be102b18/horizon-log/0.log" to get inode usage: stat /var/log/pods/openstack_horizon-754cbc7dbb-wbdt6_c176d78f-1abf-4191-94f8-8ee7be102b18/horizon-log/0.log: no such file or directory Sep 30 08:23:16 crc kubenswrapper[4810]: I0930 08:23:16.537855 4810 generic.go:334] "Generic (PLEG): container finished" podID="69265962-2e8c-430d-b982-208424d9000a" containerID="d0134ce461d7efd3f22a8c4b2cd16c9053d8b85904f7f219d708c6467c6cd8ab" exitCode=0 Sep 30 08:23:16 crc kubenswrapper[4810]: I0930 08:23:16.537928 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69265962-2e8c-430d-b982-208424d9000a","Type":"ContainerDied","Data":"d0134ce461d7efd3f22a8c4b2cd16c9053d8b85904f7f219d708c6467c6cd8ab"} Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.063850 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.197004 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.210275 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgg52\" (UniqueName: \"kubernetes.io/projected/b753b118-389f-432d-a3a8-366ebb8b440d-kube-api-access-kgg52\") pod \"b753b118-389f-432d-a3a8-366ebb8b440d\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.210342 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-log-httpd\") pod \"b753b118-389f-432d-a3a8-366ebb8b440d\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.210412 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-sg-core-conf-yaml\") pod \"b753b118-389f-432d-a3a8-366ebb8b440d\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.210486 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-combined-ca-bundle\") pod \"b753b118-389f-432d-a3a8-366ebb8b440d\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.210534 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-config-data\") pod \"b753b118-389f-432d-a3a8-366ebb8b440d\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.210631 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-scripts\") pod \"b753b118-389f-432d-a3a8-366ebb8b440d\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.210969 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b753b118-389f-432d-a3a8-366ebb8b440d" (UID: "b753b118-389f-432d-a3a8-366ebb8b440d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.211165 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-run-httpd\") pod \"b753b118-389f-432d-a3a8-366ebb8b440d\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.211426 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b753b118-389f-432d-a3a8-366ebb8b440d" (UID: "b753b118-389f-432d-a3a8-366ebb8b440d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.212024 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.212044 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b753b118-389f-432d-a3a8-366ebb8b440d-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.220311 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b753b118-389f-432d-a3a8-366ebb8b440d-kube-api-access-kgg52" (OuterVolumeSpecName: "kube-api-access-kgg52") pod "b753b118-389f-432d-a3a8-366ebb8b440d" (UID: "b753b118-389f-432d-a3a8-366ebb8b440d"). InnerVolumeSpecName "kube-api-access-kgg52". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.229145 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-scripts" (OuterVolumeSpecName: "scripts") pod "b753b118-389f-432d-a3a8-366ebb8b440d" (UID: "b753b118-389f-432d-a3a8-366ebb8b440d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.314127 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b753b118-389f-432d-a3a8-366ebb8b440d" (UID: "b753b118-389f-432d-a3a8-366ebb8b440d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.316946 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-combined-ca-bundle\") pod \"69265962-2e8c-430d-b982-208424d9000a\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.317036 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-httpd-run\") pod \"69265962-2e8c-430d-b982-208424d9000a\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.317160 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-internal-tls-certs\") pod \"69265962-2e8c-430d-b982-208424d9000a\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.317220 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-config-data\") pod \"69265962-2e8c-430d-b982-208424d9000a\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.317240 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-logs\") pod \"69265962-2e8c-430d-b982-208424d9000a\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.317295 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq67h\" (UniqueName: \"kubernetes.io/projected/69265962-2e8c-430d-b982-208424d9000a-kube-api-access-cq67h\") pod \"69265962-2e8c-430d-b982-208424d9000a\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.317315 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-scripts\") pod \"69265962-2e8c-430d-b982-208424d9000a\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.317333 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"69265962-2e8c-430d-b982-208424d9000a\" (UID: \"69265962-2e8c-430d-b982-208424d9000a\") " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.317690 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.317709 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgg52\" (UniqueName: \"kubernetes.io/projected/b753b118-389f-432d-a3a8-366ebb8b440d-kube-api-access-kgg52\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.317720 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.323567 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "69265962-2e8c-430d-b982-208424d9000a" (UID: "69265962-2e8c-430d-b982-208424d9000a"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.324517 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "69265962-2e8c-430d-b982-208424d9000a" (UID: "69265962-2e8c-430d-b982-208424d9000a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.324733 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-logs" (OuterVolumeSpecName: "logs") pod "69265962-2e8c-430d-b982-208424d9000a" (UID: "69265962-2e8c-430d-b982-208424d9000a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.335295 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-scripts" (OuterVolumeSpecName: "scripts") pod "69265962-2e8c-430d-b982-208424d9000a" (UID: "69265962-2e8c-430d-b982-208424d9000a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.335640 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69265962-2e8c-430d-b982-208424d9000a-kube-api-access-cq67h" (OuterVolumeSpecName: "kube-api-access-cq67h") pod "69265962-2e8c-430d-b982-208424d9000a" (UID: "69265962-2e8c-430d-b982-208424d9000a"). InnerVolumeSpecName "kube-api-access-cq67h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.381691 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-config-data" (OuterVolumeSpecName: "config-data") pod "69265962-2e8c-430d-b982-208424d9000a" (UID: "69265962-2e8c-430d-b982-208424d9000a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.385462 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69265962-2e8c-430d-b982-208424d9000a" (UID: "69265962-2e8c-430d-b982-208424d9000a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.413098 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b753b118-389f-432d-a3a8-366ebb8b440d" (UID: "b753b118-389f-432d-a3a8-366ebb8b440d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.418489 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-config-data" (OuterVolumeSpecName: "config-data") pod "b753b118-389f-432d-a3a8-366ebb8b440d" (UID: "b753b118-389f-432d-a3a8-366ebb8b440d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.419099 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-config-data\") pod \"b753b118-389f-432d-a3a8-366ebb8b440d\" (UID: \"b753b118-389f-432d-a3a8-366ebb8b440d\") " Sep 30 08:23:17 crc kubenswrapper[4810]: W0930 08:23:17.419203 4810 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/b753b118-389f-432d-a3a8-366ebb8b440d/volumes/kubernetes.io~secret/config-data Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.419215 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-config-data" (OuterVolumeSpecName: "config-data") pod "b753b118-389f-432d-a3a8-366ebb8b440d" (UID: "b753b118-389f-432d-a3a8-366ebb8b440d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.419731 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.419747 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.419755 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq67h\" (UniqueName: \"kubernetes.io/projected/69265962-2e8c-430d-b982-208424d9000a-kube-api-access-cq67h\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.419764 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.419784 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.419793 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.419800 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69265962-2e8c-430d-b982-208424d9000a-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.419809 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.419833 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b753b118-389f-432d-a3a8-366ebb8b440d-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.438252 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "69265962-2e8c-430d-b982-208424d9000a" (UID: "69265962-2e8c-430d-b982-208424d9000a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.442138 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.522144 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69265962-2e8c-430d-b982-208424d9000a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.522440 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.553034 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69265962-2e8c-430d-b982-208424d9000a","Type":"ContainerDied","Data":"64bbe17298c76de77325ccf3362517570f2acc52476c8b7170f764c5724d1626"} Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.553103 4810 scope.go:117] "RemoveContainer" containerID="d0134ce461d7efd3f22a8c4b2cd16c9053d8b85904f7f219d708c6467c6cd8ab" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.553259 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.558213 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f0aa786d-3e79-423b-82d1-5e9c70ca7c72","Type":"ContainerStarted","Data":"3de43c740973d856cfe9fd721f91516473195e59917ccc9bdcc591e89dcd2138"} Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.578047 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b753b118-389f-432d-a3a8-366ebb8b440d","Type":"ContainerDied","Data":"7ea36990f1e8baa8c4bd6d19e5de5041cd332ee987fa1bbf4752442e744c4581"} Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.578156 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.614900 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.306025095 podStartE2EDuration="15.614873428s" podCreationTimestamp="2025-09-30 08:23:02 +0000 UTC" firstStartedPulling="2025-09-30 08:23:03.500756687 +0000 UTC m=+1206.952955954" lastFinishedPulling="2025-09-30 08:23:16.80960502 +0000 UTC m=+1220.261804287" observedRunningTime="2025-09-30 08:23:17.5894649 +0000 UTC m=+1221.041664177" watchObservedRunningTime="2025-09-30 08:23:17.614873428 +0000 UTC m=+1221.067072695" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.626344 4810 scope.go:117] "RemoveContainer" containerID="c8f8bd13f67c179c83c43fbc5a354363acb5d744b1cc57da1a3f9aa96adc0451" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.659941 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerStarted","Data":"ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846"} Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.673596 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.684015 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.693245 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:23:17 crc kubenswrapper[4810]: E0930 08:23:17.693875 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="ceilometer-notification-agent" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.693896 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="ceilometer-notification-agent" Sep 30 08:23:17 crc kubenswrapper[4810]: E0930 08:23:17.693911 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69265962-2e8c-430d-b982-208424d9000a" containerName="glance-httpd" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.693917 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="69265962-2e8c-430d-b982-208424d9000a" containerName="glance-httpd" Sep 30 08:23:17 crc kubenswrapper[4810]: E0930 08:23:17.693947 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69265962-2e8c-430d-b982-208424d9000a" containerName="glance-log" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.693953 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="69265962-2e8c-430d-b982-208424d9000a" containerName="glance-log" Sep 30 08:23:17 crc kubenswrapper[4810]: E0930 08:23:17.693971 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="sg-core" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.693979 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="sg-core" Sep 30 08:23:17 crc kubenswrapper[4810]: E0930 08:23:17.693991 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="ceilometer-central-agent" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.693997 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="ceilometer-central-agent" Sep 30 08:23:17 crc kubenswrapper[4810]: E0930 08:23:17.694017 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="proxy-httpd" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.694022 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="proxy-httpd" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.694355 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="sg-core" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.694372 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="proxy-httpd" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.694393 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="ceilometer-notification-agent" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.694417 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="69265962-2e8c-430d-b982-208424d9000a" containerName="glance-log" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.694438 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="69265962-2e8c-430d-b982-208424d9000a" containerName="glance-httpd" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.694453 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" containerName="ceilometer-central-agent" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.695819 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.709190 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.709668 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.724171 4810 scope.go:117] "RemoveContainer" containerID="8c33afeca8d86afbe1d3c6f1b99b89e9f56ac12d2cabde574a7cbf37303b540a" Sep 30 08:23:17 crc kubenswrapper[4810]: E0930 08:23:17.767305 4810 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/9c458d0895d8803fb55a25f57f6fd96cd09100829168b99972e14f20ff1ce926/diff" to get inode usage: stat /var/lib/containers/storage/overlay/9c458d0895d8803fb55a25f57f6fd96cd09100829168b99972e14f20ff1ce926/diff: no such file or directory, extraDiskErr: Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.781369 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.797336 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.804312 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.833179 4810 scope.go:117] "RemoveContainer" containerID="e4178f188246c0cd2bc4f1b754e4d5c857a4a13c3ff2099826b5c21b847de1ea" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.838003 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.840668 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.844353 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.844555 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.855434 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.855471 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-scripts\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.855525 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sm8k\" (UniqueName: \"kubernetes.io/projected/77634c10-6e69-4f6c-9090-be8a59588745-kube-api-access-9sm8k\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.855572 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77634c10-6e69-4f6c-9090-be8a59588745-logs\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.855606 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77634c10-6e69-4f6c-9090-be8a59588745-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.855621 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-config-data\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.855676 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.855706 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.873005 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.878491 4810 scope.go:117] "RemoveContainer" containerID="a4b335d34bdb1761d899893ee878fe45cb61a29214a0a773daeea14ea0a7867b" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.908224 4810 scope.go:117] "RemoveContainer" containerID="d0dc8d2cd558ebba2195fb6632f367d95d8e2943e796177366b8402974153dd9" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.957583 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.957654 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.957689 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-config-data\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.957721 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.957746 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcqzf\" (UniqueName: \"kubernetes.io/projected/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-kube-api-access-mcqzf\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.957797 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-scripts\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.957893 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.957924 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-scripts\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.958029 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sm8k\" (UniqueName: \"kubernetes.io/projected/77634c10-6e69-4f6c-9090-be8a59588745-kube-api-access-9sm8k\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.958100 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-log-httpd\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.958127 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77634c10-6e69-4f6c-9090-be8a59588745-logs\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.958200 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77634c10-6e69-4f6c-9090-be8a59588745-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.958230 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.958254 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-config-data\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.958383 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-run-httpd\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.958979 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77634c10-6e69-4f6c-9090-be8a59588745-logs\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.959015 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77634c10-6e69-4f6c-9090-be8a59588745-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.959894 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.963450 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.964755 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.965597 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-scripts\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.970372 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77634c10-6e69-4f6c-9090-be8a59588745-config-data\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:17 crc kubenswrapper[4810]: I0930 08:23:17.980035 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sm8k\" (UniqueName: \"kubernetes.io/projected/77634c10-6e69-4f6c-9090-be8a59588745-kube-api-access-9sm8k\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.008720 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"77634c10-6e69-4f6c-9090-be8a59588745\") " pod="openstack/glance-default-internal-api-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.061004 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-log-httpd\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.061079 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.061145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-run-httpd\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.061189 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-config-data\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.061213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.061235 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcqzf\" (UniqueName: \"kubernetes.io/projected/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-kube-api-access-mcqzf\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.061271 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-scripts\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.062020 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-run-httpd\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.062081 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-log-httpd\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.067100 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.067551 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-scripts\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.071660 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.074260 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-config-data\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.079870 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.085939 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcqzf\" (UniqueName: \"kubernetes.io/projected/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-kube-api-access-mcqzf\") pod \"ceilometer-0\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.181803 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.643106 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.696304 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"77634c10-6e69-4f6c-9090-be8a59588745","Type":"ContainerStarted","Data":"5240b229a130b2ecc648ab055c0113a9d070925b30f04e2446581e055b5d2cdf"} Sep 30 08:23:18 crc kubenswrapper[4810]: I0930 08:23:18.763288 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:18 crc kubenswrapper[4810]: E0930 08:23:18.826240 4810 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/24131e76f2040b93c4194e7a8db8a8939ee213d5fa83a7c7d6146b733a45d188/diff" to get inode usage: stat /var/lib/containers/storage/overlay/24131e76f2040b93c4194e7a8db8a8939ee213d5fa83a7c7d6146b733a45d188/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_dnsmasq-dns-675956976c-gd7bn_06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b/dnsmasq-dns/0.log" to get inode usage: stat /var/log/pods/openstack_dnsmasq-dns-675956976c-gd7bn_06b5e7c0-1bd5-44fd-9ee9-4665bd93d06b/dnsmasq-dns/0.log: no such file or directory Sep 30 08:23:19 crc kubenswrapper[4810]: I0930 08:23:19.136582 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:19 crc kubenswrapper[4810]: I0930 08:23:19.148139 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5cd556568c-9c6km" Sep 30 08:23:19 crc kubenswrapper[4810]: I0930 08:23:19.323222 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69265962-2e8c-430d-b982-208424d9000a" path="/var/lib/kubelet/pods/69265962-2e8c-430d-b982-208424d9000a/volumes" Sep 30 08:23:19 crc kubenswrapper[4810]: I0930 08:23:19.324180 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b753b118-389f-432d-a3a8-366ebb8b440d" path="/var/lib/kubelet/pods/b753b118-389f-432d-a3a8-366ebb8b440d/volumes" Sep 30 08:23:19 crc kubenswrapper[4810]: I0930 08:23:19.719075 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6","Type":"ContainerStarted","Data":"05430078a9bc9b02ffe5aacdf8ef0ee9a30f1799dbe29d669bef20a22dc976b9"} Sep 30 08:23:19 crc kubenswrapper[4810]: I0930 08:23:19.719326 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6","Type":"ContainerStarted","Data":"d4266fd641bf1a603e582aaca29f7ce860ad838cfcb5914aaa03e3c2df8f04d4"} Sep 30 08:23:19 crc kubenswrapper[4810]: I0930 08:23:19.719337 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6","Type":"ContainerStarted","Data":"fa4712e97744e258fdb5f20dbd1c11df422da19c0b664317131cbf37d9caa5e8"} Sep 30 08:23:19 crc kubenswrapper[4810]: I0930 08:23:19.722435 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"77634c10-6e69-4f6c-9090-be8a59588745","Type":"ContainerStarted","Data":"75d96492e2964f46b63c566adf306e600a33322a76357f3d437919817c7a3f7c"} Sep 30 08:23:20 crc kubenswrapper[4810]: I0930 08:23:20.735183 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"77634c10-6e69-4f6c-9090-be8a59588745","Type":"ContainerStarted","Data":"7b0c73259adbfeb8dca1ce873d4c5d4962fd8173c2ee56e873118c5901463be1"} Sep 30 08:23:20 crc kubenswrapper[4810]: I0930 08:23:20.738089 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6","Type":"ContainerStarted","Data":"60d5629f4914915efb68d46749ac1fc3d52d2ed1cce74d9a1659950db962609d"} Sep 30 08:23:21 crc kubenswrapper[4810]: I0930 08:23:21.751193 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6","Type":"ContainerStarted","Data":"cd8714b5531e9fdc3af0f9648774610c226f0309526731b328d2988bef6ec797"} Sep 30 08:23:21 crc kubenswrapper[4810]: I0930 08:23:21.751975 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 08:23:21 crc kubenswrapper[4810]: I0930 08:23:21.754034 4810 generic.go:334] "Generic (PLEG): container finished" podID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerID="ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846" exitCode=1 Sep 30 08:23:21 crc kubenswrapper[4810]: I0930 08:23:21.754709 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerDied","Data":"ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846"} Sep 30 08:23:21 crc kubenswrapper[4810]: I0930 08:23:21.754747 4810 scope.go:117] "RemoveContainer" containerID="97265a9c186e6460557751232f731c5bc9eac539bb43e0580b1ec02387f960fc" Sep 30 08:23:21 crc kubenswrapper[4810]: I0930 08:23:21.755109 4810 scope.go:117] "RemoveContainer" containerID="ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846" Sep 30 08:23:21 crc kubenswrapper[4810]: E0930 08:23:21.755350 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:23:21 crc kubenswrapper[4810]: I0930 08:23:21.778772 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.778748898 podStartE2EDuration="4.778748898s" podCreationTimestamp="2025-09-30 08:23:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:23:20.768485302 +0000 UTC m=+1224.220684559" watchObservedRunningTime="2025-09-30 08:23:21.778748898 +0000 UTC m=+1225.230948165" Sep 30 08:23:21 crc kubenswrapper[4810]: I0930 08:23:21.780395 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.1734824169999998 podStartE2EDuration="4.780386425s" podCreationTimestamp="2025-09-30 08:23:17 +0000 UTC" firstStartedPulling="2025-09-30 08:23:18.770072765 +0000 UTC m=+1222.222272032" lastFinishedPulling="2025-09-30 08:23:21.376976753 +0000 UTC m=+1224.829176040" observedRunningTime="2025-09-30 08:23:21.77434218 +0000 UTC m=+1225.226541477" watchObservedRunningTime="2025-09-30 08:23:21.780386425 +0000 UTC m=+1225.232585692" Sep 30 08:23:23 crc kubenswrapper[4810]: I0930 08:23:23.479064 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:23:23 crc kubenswrapper[4810]: I0930 08:23:23.480552 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:23:23 crc kubenswrapper[4810]: I0930 08:23:23.481249 4810 scope.go:117] "RemoveContainer" containerID="ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846" Sep 30 08:23:23 crc kubenswrapper[4810]: E0930 08:23:23.481556 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:23:23 crc kubenswrapper[4810]: I0930 08:23:23.644845 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:23:23 crc kubenswrapper[4810]: I0930 08:23:23.645200 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="852f98b7-5c49-4876-9530-1bacda996d61" containerName="glance-log" containerID="cri-o://6c5fdae1523b62cd2c8f8055eb375f1729321df4f5a89cbd08c1074f4cac39f4" gracePeriod=30 Sep 30 08:23:23 crc kubenswrapper[4810]: I0930 08:23:23.645316 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="852f98b7-5c49-4876-9530-1bacda996d61" containerName="glance-httpd" containerID="cri-o://ce6ef8f3ade002003120fee17fee20ef6302aebca084e8c99dcb4fe6d1102e3b" gracePeriod=30 Sep 30 08:23:23 crc kubenswrapper[4810]: I0930 08:23:23.784964 4810 generic.go:334] "Generic (PLEG): container finished" podID="852f98b7-5c49-4876-9530-1bacda996d61" containerID="6c5fdae1523b62cd2c8f8055eb375f1729321df4f5a89cbd08c1074f4cac39f4" exitCode=143 Sep 30 08:23:23 crc kubenswrapper[4810]: I0930 08:23:23.786135 4810 scope.go:117] "RemoveContainer" containerID="ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846" Sep 30 08:23:23 crc kubenswrapper[4810]: E0930 08:23:23.786420 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:23:23 crc kubenswrapper[4810]: I0930 08:23:23.786745 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"852f98b7-5c49-4876-9530-1bacda996d61","Type":"ContainerDied","Data":"6c5fdae1523b62cd2c8f8055eb375f1729321df4f5a89cbd08c1074f4cac39f4"} Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.428420 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.492751 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-scripts\") pod \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.492861 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-etc-machine-id\") pod \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.492923 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data\") pod \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.492978 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-combined-ca-bundle\") pod \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.493015 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data-custom\") pod \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.493194 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgb2w\" (UniqueName: \"kubernetes.io/projected/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-kube-api-access-pgb2w\") pod \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.493265 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-logs\") pod \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\" (UID: \"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b\") " Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.493864 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" (UID: "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.494141 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-logs" (OuterVolumeSpecName: "logs") pod "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" (UID: "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.501431 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" (UID: "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.505297 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-kube-api-access-pgb2w" (OuterVolumeSpecName: "kube-api-access-pgb2w") pod "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" (UID: "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b"). InnerVolumeSpecName "kube-api-access-pgb2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.505463 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-scripts" (OuterVolumeSpecName: "scripts") pod "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" (UID: "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:24 crc kubenswrapper[4810]: W0930 08:23:24.563344 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb753b118_389f_432d_a3a8_366ebb8b440d.slice/crio-8c33afeca8d86afbe1d3c6f1b99b89e9f56ac12d2cabde574a7cbf37303b540a.scope WatchSource:0}: Error finding container 8c33afeca8d86afbe1d3c6f1b99b89e9f56ac12d2cabde574a7cbf37303b540a: Status 404 returned error can't find the container with id 8c33afeca8d86afbe1d3c6f1b99b89e9f56ac12d2cabde574a7cbf37303b540a Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.571522 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" (UID: "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:24 crc kubenswrapper[4810]: W0930 08:23:24.593816 4810 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6acc6a2_b3e3_4a1a_8e4e_a0e8fff58892.slice/crio-conmon-ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6acc6a2_b3e3_4a1a_8e4e_a0e8fff58892.slice/crio-conmon-ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846.scope: no such file or directory Sep 30 08:23:24 crc kubenswrapper[4810]: W0930 08:23:24.593892 4810 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6acc6a2_b3e3_4a1a_8e4e_a0e8fff58892.slice/crio-ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6acc6a2_b3e3_4a1a_8e4e_a0e8fff58892.slice/crio-ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846.scope: no such file or directory Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.595564 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.595593 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.595605 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.595619 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgb2w\" (UniqueName: \"kubernetes.io/projected/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-kube-api-access-pgb2w\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.595632 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.595643 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.611099 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data" (OuterVolumeSpecName: "config-data") pod "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" (UID: "79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.697028 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.803628 4810 generic.go:334] "Generic (PLEG): container finished" podID="852f98b7-5c49-4876-9530-1bacda996d61" containerID="ce6ef8f3ade002003120fee17fee20ef6302aebca084e8c99dcb4fe6d1102e3b" exitCode=0 Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.803916 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"852f98b7-5c49-4876-9530-1bacda996d61","Type":"ContainerDied","Data":"ce6ef8f3ade002003120fee17fee20ef6302aebca084e8c99dcb4fe6d1102e3b"} Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.819424 4810 generic.go:334] "Generic (PLEG): container finished" podID="79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" containerID="a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b" exitCode=137 Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.819508 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b","Type":"ContainerDied","Data":"a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b"} Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.819570 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b","Type":"ContainerDied","Data":"d9fba371e8c661458905bde13739c6e7f14605eb84de0e64580b79ff49ee28d6"} Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.819598 4810 scope.go:117] "RemoveContainer" containerID="a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.819825 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 08:23:24 crc kubenswrapper[4810]: E0930 08:23:24.851889 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod852f98b7_5c49_4876_9530_1bacda996d61.slice/crio-ce6ef8f3ade002003120fee17fee20ef6302aebca084e8c99dcb4fe6d1102e3b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod852f98b7_5c49_4876_9530_1bacda996d61.slice/crio-conmon-ce6ef8f3ade002003120fee17fee20ef6302aebca084e8c99dcb4fe6d1102e3b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69265962_2e8c_430d_b982_208424d9000a.slice/crio-64bbe17298c76de77325ccf3362517570f2acc52476c8b7170f764c5724d1626\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb753b118_389f_432d_a3a8_366ebb8b440d.slice/crio-d0dc8d2cd558ebba2195fb6632f367d95d8e2943e796177366b8402974153dd9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ac68abd_cf5c_4a46_9db1_599d547c3b06.slice/crio-conmon-fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb753b118_389f_432d_a3a8_366ebb8b440d.slice/crio-7ea36990f1e8baa8c4bd6d19e5de5041cd332ee987fa1bbf4752442e744c4581\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ac68abd_cf5c_4a46_9db1_599d547c3b06.slice/crio-conmon-cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ac68abd_cf5c_4a46_9db1_599d547c3b06.slice/crio-730c6d8f1bd112a78dda99f986a7691880de092ca2d81faef317b0f3a0882db4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod852f98b7_5c49_4876_9530_1bacda996d61.slice/crio-conmon-6c5fdae1523b62cd2c8f8055eb375f1729321df4f5a89cbd08c1074f4cac39f4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69265962_2e8c_430d_b982_208424d9000a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb753b118_389f_432d_a3a8_366ebb8b440d.slice/crio-conmon-a4b335d34bdb1761d899893ee878fe45cb61a29214a0a773daeea14ea0a7867b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69265962_2e8c_430d_b982_208424d9000a.slice/crio-d0134ce461d7efd3f22a8c4b2cd16c9053d8b85904f7f219d708c6467c6cd8ab.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod852f98b7_5c49_4876_9530_1bacda996d61.slice/crio-6c5fdae1523b62cd2c8f8055eb375f1729321df4f5a89cbd08c1074f4cac39f4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ac68abd_cf5c_4a46_9db1_599d547c3b06.slice/crio-fb6a8f323f6f581751adf739f9e979cf5783d738ddbccb0678a0a9c6d5e9a351.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb753b118_389f_432d_a3a8_366ebb8b440d.slice/crio-conmon-d0dc8d2cd558ebba2195fb6632f367d95d8e2943e796177366b8402974153dd9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc176d78f_1abf_4191_94f8_8ee7be102b18.slice/crio-65c41b5d112da8a316221d00dc0b08a61accf182470442b32f5aa0e475f2698f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb753b118_389f_432d_a3a8_366ebb8b440d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ac68abd_cf5c_4a46_9db1_599d547c3b06.slice/crio-cf129000c878f40d100136300ed2941d5b8ef80ab65c125cb4167d2d4cd5c491.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc176d78f_1abf_4191_94f8_8ee7be102b18.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb753b118_389f_432d_a3a8_366ebb8b440d.slice/crio-a4b335d34bdb1761d899893ee878fe45cb61a29214a0a773daeea14ea0a7867b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb753b118_389f_432d_a3a8_366ebb8b440d.slice/crio-conmon-8c33afeca8d86afbe1d3c6f1b99b89e9f56ac12d2cabde574a7cbf37303b540a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69265962_2e8c_430d_b982_208424d9000a.slice/crio-c8f8bd13f67c179c83c43fbc5a354363acb5d744b1cc57da1a3f9aa96adc0451.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69265962_2e8c_430d_b982_208424d9000a.slice/crio-conmon-c8f8bd13f67c179c83c43fbc5a354363acb5d744b1cc57da1a3f9aa96adc0451.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69265962_2e8c_430d_b982_208424d9000a.slice/crio-conmon-d0134ce461d7efd3f22a8c4b2cd16c9053d8b85904f7f219d708c6467c6cd8ab.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc176d78f_1abf_4191_94f8_8ee7be102b18.slice/crio-conmon-65c41b5d112da8a316221d00dc0b08a61accf182470442b32f5aa0e475f2698f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb753b118_389f_432d_a3a8_366ebb8b440d.slice/crio-conmon-e4178f188246c0cd2bc4f1b754e4d5c857a4a13c3ff2099826b5c21b847de1ea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc176d78f_1abf_4191_94f8_8ee7be102b18.slice/crio-c4a402a0d631408a6d207e0bd44625463f0d1368dd2485a3e052ac7d06864225\": RecentStats: unable to find data in memory cache]" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.859481 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.860276 4810 scope.go:117] "RemoveContainer" containerID="5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.866929 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.907005 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 30 08:23:24 crc kubenswrapper[4810]: E0930 08:23:24.907420 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" containerName="cinder-api-log" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.907433 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" containerName="cinder-api-log" Sep 30 08:23:24 crc kubenswrapper[4810]: E0930 08:23:24.907480 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" containerName="cinder-api" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.907487 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" containerName="cinder-api" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.907843 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" containerName="cinder-api-log" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.907872 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" containerName="cinder-api" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.908873 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.913606 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.913747 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.914114 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.920918 4810 scope.go:117] "RemoveContainer" containerID="a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b" Sep 30 08:23:24 crc kubenswrapper[4810]: E0930 08:23:24.921647 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b\": container with ID starting with a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b not found: ID does not exist" containerID="a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.921681 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b"} err="failed to get container status \"a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b\": rpc error: code = NotFound desc = could not find container \"a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b\": container with ID starting with a093f5e4301a9b2f19a4459f215eb2f94540a1837aa3c8bf453bdbdf1e62d01b not found: ID does not exist" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.921753 4810 scope.go:117] "RemoveContainer" containerID="5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348" Sep 30 08:23:24 crc kubenswrapper[4810]: E0930 08:23:24.922049 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348\": container with ID starting with 5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348 not found: ID does not exist" containerID="5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.922089 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348"} err="failed to get container status \"5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348\": rpc error: code = NotFound desc = could not find container \"5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348\": container with ID starting with 5d749151617b3104ef59250164ee150b0fbac5cc71d34688eb8ef575136eb348 not found: ID does not exist" Sep 30 08:23:24 crc kubenswrapper[4810]: I0930 08:23:24.934320 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.002053 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.002100 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-scripts\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.002128 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.002152 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58537813-eea7-4a58-a900-f3473797d7d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.002283 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58537813-eea7-4a58-a900-f3473797d7d2-logs\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.002300 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.002320 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q52pj\" (UniqueName: \"kubernetes.io/projected/58537813-eea7-4a58-a900-f3473797d7d2-kube-api-access-q52pj\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.002355 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.002485 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-config-data\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.066504 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.066808 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="ceilometer-central-agent" containerID="cri-o://d4266fd641bf1a603e582aaca29f7ce860ad838cfcb5914aaa03e3c2df8f04d4" gracePeriod=30 Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.067340 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="proxy-httpd" containerID="cri-o://cd8714b5531e9fdc3af0f9648774610c226f0309526731b328d2988bef6ec797" gracePeriod=30 Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.067393 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="sg-core" containerID="cri-o://60d5629f4914915efb68d46749ac1fc3d52d2ed1cce74d9a1659950db962609d" gracePeriod=30 Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.067427 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="ceilometer-notification-agent" containerID="cri-o://05430078a9bc9b02ffe5aacdf8ef0ee9a30f1799dbe29d669bef20a22dc976b9" gracePeriod=30 Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.104582 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58537813-eea7-4a58-a900-f3473797d7d2-logs\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.104633 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.104654 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q52pj\" (UniqueName: \"kubernetes.io/projected/58537813-eea7-4a58-a900-f3473797d7d2-kube-api-access-q52pj\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.104691 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.104725 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-config-data\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.104784 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.104812 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-scripts\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.104838 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.104860 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58537813-eea7-4a58-a900-f3473797d7d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.104969 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58537813-eea7-4a58-a900-f3473797d7d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.105838 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58537813-eea7-4a58-a900-f3473797d7d2-logs\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.111184 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-scripts\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.111483 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.111799 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.112586 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-config-data\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.113029 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.114130 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58537813-eea7-4a58-a900-f3473797d7d2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.125953 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q52pj\" (UniqueName: \"kubernetes.io/projected/58537813-eea7-4a58-a900-f3473797d7d2-kube-api-access-q52pj\") pod \"cinder-api-0\" (UID: \"58537813-eea7-4a58-a900-f3473797d7d2\") " pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.209911 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.240450 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.307088 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-logs\") pod \"852f98b7-5c49-4876-9530-1bacda996d61\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.307212 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-scripts\") pod \"852f98b7-5c49-4876-9530-1bacda996d61\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.307244 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-public-tls-certs\") pod \"852f98b7-5c49-4876-9530-1bacda996d61\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.307347 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-httpd-run\") pod \"852f98b7-5c49-4876-9530-1bacda996d61\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.307400 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-combined-ca-bundle\") pod \"852f98b7-5c49-4876-9530-1bacda996d61\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.307461 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"852f98b7-5c49-4876-9530-1bacda996d61\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.307545 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn9vd\" (UniqueName: \"kubernetes.io/projected/852f98b7-5c49-4876-9530-1bacda996d61-kube-api-access-fn9vd\") pod \"852f98b7-5c49-4876-9530-1bacda996d61\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.307622 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-config-data\") pod \"852f98b7-5c49-4876-9530-1bacda996d61\" (UID: \"852f98b7-5c49-4876-9530-1bacda996d61\") " Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.307963 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "852f98b7-5c49-4876-9530-1bacda996d61" (UID: "852f98b7-5c49-4876-9530-1bacda996d61"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.308088 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-logs" (OuterVolumeSpecName: "logs") pod "852f98b7-5c49-4876-9530-1bacda996d61" (UID: "852f98b7-5c49-4876-9530-1bacda996d61"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.320664 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-scripts" (OuterVolumeSpecName: "scripts") pod "852f98b7-5c49-4876-9530-1bacda996d61" (UID: "852f98b7-5c49-4876-9530-1bacda996d61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.323181 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "852f98b7-5c49-4876-9530-1bacda996d61" (UID: "852f98b7-5c49-4876-9530-1bacda996d61"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.323877 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/852f98b7-5c49-4876-9530-1bacda996d61-kube-api-access-fn9vd" (OuterVolumeSpecName: "kube-api-access-fn9vd") pod "852f98b7-5c49-4876-9530-1bacda996d61" (UID: "852f98b7-5c49-4876-9530-1bacda996d61"). InnerVolumeSpecName "kube-api-access-fn9vd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.327475 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b" path="/var/lib/kubelet/pods/79ec9ff9-6c95-4fa1-86d4-e6e028b8e02b/volumes" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.353943 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "852f98b7-5c49-4876-9530-1bacda996d61" (UID: "852f98b7-5c49-4876-9530-1bacda996d61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.376488 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-config-data" (OuterVolumeSpecName: "config-data") pod "852f98b7-5c49-4876-9530-1bacda996d61" (UID: "852f98b7-5c49-4876-9530-1bacda996d61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.391532 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "852f98b7-5c49-4876-9530-1bacda996d61" (UID: "852f98b7-5c49-4876-9530-1bacda996d61"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.412709 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.412739 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn9vd\" (UniqueName: \"kubernetes.io/projected/852f98b7-5c49-4876-9530-1bacda996d61-kube-api-access-fn9vd\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.412750 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.412757 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.412765 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.412777 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.412786 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/852f98b7-5c49-4876-9530-1bacda996d61-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.412795 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/852f98b7-5c49-4876-9530-1bacda996d61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.437132 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.513804 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.763114 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.846621 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.846685 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"852f98b7-5c49-4876-9530-1bacda996d61","Type":"ContainerDied","Data":"d843917bbb0111c02f83752026cec010eefe565c5adde5ffb169c8dec33046f0"} Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.846732 4810 scope.go:117] "RemoveContainer" containerID="ce6ef8f3ade002003120fee17fee20ef6302aebca084e8c99dcb4fe6d1102e3b" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.862217 4810 generic.go:334] "Generic (PLEG): container finished" podID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerID="cd8714b5531e9fdc3af0f9648774610c226f0309526731b328d2988bef6ec797" exitCode=0 Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.862307 4810 generic.go:334] "Generic (PLEG): container finished" podID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerID="60d5629f4914915efb68d46749ac1fc3d52d2ed1cce74d9a1659950db962609d" exitCode=2 Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.862317 4810 generic.go:334] "Generic (PLEG): container finished" podID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerID="05430078a9bc9b02ffe5aacdf8ef0ee9a30f1799dbe29d669bef20a22dc976b9" exitCode=0 Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.862325 4810 generic.go:334] "Generic (PLEG): container finished" podID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerID="d4266fd641bf1a603e582aaca29f7ce860ad838cfcb5914aaa03e3c2df8f04d4" exitCode=0 Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.862564 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6","Type":"ContainerDied","Data":"cd8714b5531e9fdc3af0f9648774610c226f0309526731b328d2988bef6ec797"} Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.862602 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6","Type":"ContainerDied","Data":"60d5629f4914915efb68d46749ac1fc3d52d2ed1cce74d9a1659950db962609d"} Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.862615 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6","Type":"ContainerDied","Data":"05430078a9bc9b02ffe5aacdf8ef0ee9a30f1799dbe29d669bef20a22dc976b9"} Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.862647 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6","Type":"ContainerDied","Data":"d4266fd641bf1a603e582aaca29f7ce860ad838cfcb5914aaa03e3c2df8f04d4"} Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.864098 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58537813-eea7-4a58-a900-f3473797d7d2","Type":"ContainerStarted","Data":"061dbb5042869c42b9982ba7ca0d5891ce329eef0b48e099b06816ea663877c9"} Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.913385 4810 scope.go:117] "RemoveContainer" containerID="6c5fdae1523b62cd2c8f8055eb375f1729321df4f5a89cbd08c1074f4cac39f4" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.924978 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.933556 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.949279 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:23:25 crc kubenswrapper[4810]: E0930 08:23:25.949745 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="852f98b7-5c49-4876-9530-1bacda996d61" containerName="glance-httpd" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.949762 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="852f98b7-5c49-4876-9530-1bacda996d61" containerName="glance-httpd" Sep 30 08:23:25 crc kubenswrapper[4810]: E0930 08:23:25.949787 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="852f98b7-5c49-4876-9530-1bacda996d61" containerName="glance-log" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.949794 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="852f98b7-5c49-4876-9530-1bacda996d61" containerName="glance-log" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.949979 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="852f98b7-5c49-4876-9530-1bacda996d61" containerName="glance-httpd" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.949993 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="852f98b7-5c49-4876-9530-1bacda996d61" containerName="glance-log" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.951011 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.953962 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.954024 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 08:23:25 crc kubenswrapper[4810]: I0930 08:23:25.958737 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.099482 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.126984 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-run-httpd\") pod \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.127069 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-sg-core-conf-yaml\") pod \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.127093 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-combined-ca-bundle\") pod \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.127121 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-log-httpd\") pod \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.127246 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-scripts\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.127346 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-config-data\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.127404 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.127445 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.127475 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6725\" (UniqueName: \"kubernetes.io/projected/f35de952-dcbd-422d-bbdb-b8995de342a1-kube-api-access-h6725\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.127529 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.127551 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f35de952-dcbd-422d-bbdb-b8995de342a1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.127603 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f35de952-dcbd-422d-bbdb-b8995de342a1-logs\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.128849 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" (UID: "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.129457 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" (UID: "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.173538 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" (UID: "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.229361 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-scripts\") pod \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.229548 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcqzf\" (UniqueName: \"kubernetes.io/projected/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-kube-api-access-mcqzf\") pod \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.229588 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-config-data\") pod \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\" (UID: \"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6\") " Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.229955 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.229987 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f35de952-dcbd-422d-bbdb-b8995de342a1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.230031 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f35de952-dcbd-422d-bbdb-b8995de342a1-logs\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.230066 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-scripts\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.230097 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-config-data\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.230138 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.230166 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.230192 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6725\" (UniqueName: \"kubernetes.io/projected/f35de952-dcbd-422d-bbdb-b8995de342a1-kube-api-access-h6725\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.230281 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.230293 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.230307 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.231527 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f35de952-dcbd-422d-bbdb-b8995de342a1-logs\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.233779 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.233847 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" (UID: "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.234918 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-kube-api-access-mcqzf" (OuterVolumeSpecName: "kube-api-access-mcqzf") pod "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" (UID: "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6"). InnerVolumeSpecName "kube-api-access-mcqzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.234961 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-scripts" (OuterVolumeSpecName: "scripts") pod "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" (UID: "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.235231 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f35de952-dcbd-422d-bbdb-b8995de342a1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.236294 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-config-data\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.238647 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.240449 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-scripts\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.249537 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f35de952-dcbd-422d-bbdb-b8995de342a1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.256738 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6725\" (UniqueName: \"kubernetes.io/projected/f35de952-dcbd-422d-bbdb-b8995de342a1-kube-api-access-h6725\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.271287 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f35de952-dcbd-422d-bbdb-b8995de342a1\") " pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.288407 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.333111 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.333146 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.333156 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcqzf\" (UniqueName: \"kubernetes.io/projected/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-kube-api-access-mcqzf\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.380307 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-config-data" (OuterVolumeSpecName: "config-data") pod "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" (UID: "3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.442391 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:26 crc kubenswrapper[4810]: W0930 08:23:26.886013 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf35de952_dcbd_422d_bbdb_b8995de342a1.slice/crio-b83a016c2ae4d9b47ddd5c399e5d8a41e9943a0a3efed821e865d13251ee2e18 WatchSource:0}: Error finding container b83a016c2ae4d9b47ddd5c399e5d8a41e9943a0a3efed821e865d13251ee2e18: Status 404 returned error can't find the container with id b83a016c2ae4d9b47ddd5c399e5d8a41e9943a0a3efed821e865d13251ee2e18 Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.887479 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.894401 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58537813-eea7-4a58-a900-f3473797d7d2","Type":"ContainerStarted","Data":"ef6fcad53c8a9598ff5a239453f8e38ed9c3405c3205c1de1c43865bb665a0da"} Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.915596 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6","Type":"ContainerDied","Data":"fa4712e97744e258fdb5f20dbd1c11df422da19c0b664317131cbf37d9caa5e8"} Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.915657 4810 scope.go:117] "RemoveContainer" containerID="cd8714b5531e9fdc3af0f9648774610c226f0309526731b328d2988bef6ec797" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.916313 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.975596 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:26 crc kubenswrapper[4810]: I0930 08:23:26.999833 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.004544 4810 scope.go:117] "RemoveContainer" containerID="60d5629f4914915efb68d46749ac1fc3d52d2ed1cce74d9a1659950db962609d" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.011558 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:27 crc kubenswrapper[4810]: E0930 08:23:27.012070 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="ceilometer-central-agent" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.012096 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="ceilometer-central-agent" Sep 30 08:23:27 crc kubenswrapper[4810]: E0930 08:23:27.012133 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="sg-core" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.012142 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="sg-core" Sep 30 08:23:27 crc kubenswrapper[4810]: E0930 08:23:27.012163 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="ceilometer-notification-agent" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.012171 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="ceilometer-notification-agent" Sep 30 08:23:27 crc kubenswrapper[4810]: E0930 08:23:27.012188 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="proxy-httpd" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.012194 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="proxy-httpd" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.012460 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="proxy-httpd" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.012477 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="ceilometer-central-agent" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.012497 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="sg-core" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.012513 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" containerName="ceilometer-notification-agent" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.014945 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.018007 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.018229 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.034868 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.077487 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-run-httpd\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.077552 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-scripts\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.077618 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dzvm\" (UniqueName: \"kubernetes.io/projected/bbe2f72a-f3b3-486f-8226-400a43ea161a-kube-api-access-4dzvm\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.077638 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-config-data\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.077662 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.077685 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-log-httpd\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.077750 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.091257 4810 scope.go:117] "RemoveContainer" containerID="05430078a9bc9b02ffe5aacdf8ef0ee9a30f1799dbe29d669bef20a22dc976b9" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.115945 4810 scope.go:117] "RemoveContainer" containerID="d4266fd641bf1a603e582aaca29f7ce860ad838cfcb5914aaa03e3c2df8f04d4" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.180106 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dzvm\" (UniqueName: \"kubernetes.io/projected/bbe2f72a-f3b3-486f-8226-400a43ea161a-kube-api-access-4dzvm\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.180192 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-config-data\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.180259 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.180311 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-log-httpd\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.180476 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.180558 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-run-httpd\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.180630 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-scripts\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.180824 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-log-httpd\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.181407 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-run-httpd\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.186426 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-config-data\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.188084 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.190210 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.199945 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-scripts\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.200405 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dzvm\" (UniqueName: \"kubernetes.io/projected/bbe2f72a-f3b3-486f-8226-400a43ea161a-kube-api-access-4dzvm\") pod \"ceilometer-0\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.321976 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6" path="/var/lib/kubelet/pods/3d59f6e5-bdd8-4f73-83fc-e107e83ff0d6/volumes" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.323592 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="852f98b7-5c49-4876-9530-1bacda996d61" path="/var/lib/kubelet/pods/852f98b7-5c49-4876-9530-1bacda996d61/volumes" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.379904 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.862382 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:27 crc kubenswrapper[4810]: W0930 08:23:27.872100 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbe2f72a_f3b3_486f_8226_400a43ea161a.slice/crio-eee6123203b2df6fc3230b44aaaa1bbea81de0c38ca22b557142c9da14dcfa72 WatchSource:0}: Error finding container eee6123203b2df6fc3230b44aaaa1bbea81de0c38ca22b557142c9da14dcfa72: Status 404 returned error can't find the container with id eee6123203b2df6fc3230b44aaaa1bbea81de0c38ca22b557142c9da14dcfa72 Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.930078 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58537813-eea7-4a58-a900-f3473797d7d2","Type":"ContainerStarted","Data":"b25bf81b5638eac90d4520e5a80dc61dfa654283bae392ea9e69b3d5d8cbb898"} Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.930195 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.932230 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f35de952-dcbd-422d-bbdb-b8995de342a1","Type":"ContainerStarted","Data":"6264d911f95940b20c81dde610faf44a2283e8e2e4ea797ebdc8f7fc5062f85f"} Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.932303 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f35de952-dcbd-422d-bbdb-b8995de342a1","Type":"ContainerStarted","Data":"b83a016c2ae4d9b47ddd5c399e5d8a41e9943a0a3efed821e865d13251ee2e18"} Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.935332 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe2f72a-f3b3-486f-8226-400a43ea161a","Type":"ContainerStarted","Data":"eee6123203b2df6fc3230b44aaaa1bbea81de0c38ca22b557142c9da14dcfa72"} Sep 30 08:23:27 crc kubenswrapper[4810]: I0930 08:23:27.947450 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.9474229 podStartE2EDuration="3.9474229s" podCreationTimestamp="2025-09-30 08:23:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:23:27.945800523 +0000 UTC m=+1231.397999810" watchObservedRunningTime="2025-09-30 08:23:27.9474229 +0000 UTC m=+1231.399622167" Sep 30 08:23:28 crc kubenswrapper[4810]: I0930 08:23:28.081174 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:28 crc kubenswrapper[4810]: I0930 08:23:28.081229 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:28 crc kubenswrapper[4810]: I0930 08:23:28.131307 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:28 crc kubenswrapper[4810]: I0930 08:23:28.131689 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:28 crc kubenswrapper[4810]: I0930 08:23:28.952158 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f35de952-dcbd-422d-bbdb-b8995de342a1","Type":"ContainerStarted","Data":"5cc62b0f543b8b2c76ecdce582e9f6a62344ed6812e57d58b06aa4fba4c2f2f5"} Sep 30 08:23:28 crc kubenswrapper[4810]: I0930 08:23:28.957498 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe2f72a-f3b3-486f-8226-400a43ea161a","Type":"ContainerStarted","Data":"8e9ed83b4714ff39f1383727d350ae37a3b08c7612ca43cc680dcee98117f410"} Sep 30 08:23:28 crc kubenswrapper[4810]: I0930 08:23:28.957556 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe2f72a-f3b3-486f-8226-400a43ea161a","Type":"ContainerStarted","Data":"fa1405eac25bff82b76214a60c2e7583761f1d6bb64e477c50c3991118f3ed4b"} Sep 30 08:23:28 crc kubenswrapper[4810]: I0930 08:23:28.957580 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:28 crc kubenswrapper[4810]: I0930 08:23:28.957599 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:28 crc kubenswrapper[4810]: I0930 08:23:28.977714 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.977693366 podStartE2EDuration="3.977693366s" podCreationTimestamp="2025-09-30 08:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:23:28.971805715 +0000 UTC m=+1232.424004982" watchObservedRunningTime="2025-09-30 08:23:28.977693366 +0000 UTC m=+1232.429892633" Sep 30 08:23:29 crc kubenswrapper[4810]: I0930 08:23:29.446035 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:29 crc kubenswrapper[4810]: I0930 08:23:29.971498 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe2f72a-f3b3-486f-8226-400a43ea161a","Type":"ContainerStarted","Data":"b184a13912d3c57f1ddcad66c8f7bcb029784998b921153d5e8a2672744e65a0"} Sep 30 08:23:30 crc kubenswrapper[4810]: I0930 08:23:30.837082 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:30 crc kubenswrapper[4810]: I0930 08:23:30.919941 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 08:23:30 crc kubenswrapper[4810]: I0930 08:23:30.983878 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="ceilometer-central-agent" containerID="cri-o://fa1405eac25bff82b76214a60c2e7583761f1d6bb64e477c50c3991118f3ed4b" gracePeriod=30 Sep 30 08:23:30 crc kubenswrapper[4810]: I0930 08:23:30.984169 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe2f72a-f3b3-486f-8226-400a43ea161a","Type":"ContainerStarted","Data":"afdfc89354db8734d410ed7e3f0d47c9d0b652c872bc87379955611e3aaa1a50"} Sep 30 08:23:30 crc kubenswrapper[4810]: I0930 08:23:30.984214 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 08:23:30 crc kubenswrapper[4810]: I0930 08:23:30.984495 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="proxy-httpd" containerID="cri-o://afdfc89354db8734d410ed7e3f0d47c9d0b652c872bc87379955611e3aaa1a50" gracePeriod=30 Sep 30 08:23:30 crc kubenswrapper[4810]: I0930 08:23:30.984562 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="sg-core" containerID="cri-o://b184a13912d3c57f1ddcad66c8f7bcb029784998b921153d5e8a2672744e65a0" gracePeriod=30 Sep 30 08:23:30 crc kubenswrapper[4810]: I0930 08:23:30.984604 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="ceilometer-notification-agent" containerID="cri-o://8e9ed83b4714ff39f1383727d350ae37a3b08c7612ca43cc680dcee98117f410" gracePeriod=30 Sep 30 08:23:31 crc kubenswrapper[4810]: I0930 08:23:31.019969 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.541561394 podStartE2EDuration="5.019921887s" podCreationTimestamp="2025-09-30 08:23:26 +0000 UTC" firstStartedPulling="2025-09-30 08:23:27.88994914 +0000 UTC m=+1231.342148407" lastFinishedPulling="2025-09-30 08:23:30.368309633 +0000 UTC m=+1233.820508900" observedRunningTime="2025-09-30 08:23:31.007580318 +0000 UTC m=+1234.459779585" watchObservedRunningTime="2025-09-30 08:23:31.019921887 +0000 UTC m=+1234.472121154" Sep 30 08:23:31 crc kubenswrapper[4810]: I0930 08:23:31.996546 4810 generic.go:334] "Generic (PLEG): container finished" podID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerID="afdfc89354db8734d410ed7e3f0d47c9d0b652c872bc87379955611e3aaa1a50" exitCode=0 Sep 30 08:23:31 crc kubenswrapper[4810]: I0930 08:23:31.996583 4810 generic.go:334] "Generic (PLEG): container finished" podID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerID="b184a13912d3c57f1ddcad66c8f7bcb029784998b921153d5e8a2672744e65a0" exitCode=2 Sep 30 08:23:31 crc kubenswrapper[4810]: I0930 08:23:31.996592 4810 generic.go:334] "Generic (PLEG): container finished" podID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerID="8e9ed83b4714ff39f1383727d350ae37a3b08c7612ca43cc680dcee98117f410" exitCode=0 Sep 30 08:23:31 crc kubenswrapper[4810]: I0930 08:23:31.996594 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe2f72a-f3b3-486f-8226-400a43ea161a","Type":"ContainerDied","Data":"afdfc89354db8734d410ed7e3f0d47c9d0b652c872bc87379955611e3aaa1a50"} Sep 30 08:23:31 crc kubenswrapper[4810]: I0930 08:23:31.996644 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe2f72a-f3b3-486f-8226-400a43ea161a","Type":"ContainerDied","Data":"b184a13912d3c57f1ddcad66c8f7bcb029784998b921153d5e8a2672744e65a0"} Sep 30 08:23:31 crc kubenswrapper[4810]: I0930 08:23:31.996653 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe2f72a-f3b3-486f-8226-400a43ea161a","Type":"ContainerDied","Data":"8e9ed83b4714ff39f1383727d350ae37a3b08c7612ca43cc680dcee98117f410"} Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.001435 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8q22m"] Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.002579 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8q22m" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.073483 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8q22m"] Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.145351 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-xjw8l"] Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.146800 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xjw8l" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.153469 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xjw8l"] Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.190838 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgp6c\" (UniqueName: \"kubernetes.io/projected/c40b4429-5a41-4ea8-b9ac-17bba8788f6a-kube-api-access-wgp6c\") pod \"nova-api-db-create-8q22m\" (UID: \"c40b4429-5a41-4ea8-b9ac-17bba8788f6a\") " pod="openstack/nova-api-db-create-8q22m" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.293161 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffhp2\" (UniqueName: \"kubernetes.io/projected/d61500fa-1a5e-420a-a0d9-990c98754b13-kube-api-access-ffhp2\") pod \"nova-cell0-db-create-xjw8l\" (UID: \"d61500fa-1a5e-420a-a0d9-990c98754b13\") " pod="openstack/nova-cell0-db-create-xjw8l" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.293231 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgp6c\" (UniqueName: \"kubernetes.io/projected/c40b4429-5a41-4ea8-b9ac-17bba8788f6a-kube-api-access-wgp6c\") pod \"nova-api-db-create-8q22m\" (UID: \"c40b4429-5a41-4ea8-b9ac-17bba8788f6a\") " pod="openstack/nova-api-db-create-8q22m" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.315809 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgp6c\" (UniqueName: \"kubernetes.io/projected/c40b4429-5a41-4ea8-b9ac-17bba8788f6a-kube-api-access-wgp6c\") pod \"nova-api-db-create-8q22m\" (UID: \"c40b4429-5a41-4ea8-b9ac-17bba8788f6a\") " pod="openstack/nova-api-db-create-8q22m" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.319247 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-nnzr4"] Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.321000 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nnzr4" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.350249 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-nnzr4"] Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.371081 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8q22m" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.397992 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffhp2\" (UniqueName: \"kubernetes.io/projected/d61500fa-1a5e-420a-a0d9-990c98754b13-kube-api-access-ffhp2\") pod \"nova-cell0-db-create-xjw8l\" (UID: \"d61500fa-1a5e-420a-a0d9-990c98754b13\") " pod="openstack/nova-cell0-db-create-xjw8l" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.398124 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6548\" (UniqueName: \"kubernetes.io/projected/34b10dbc-a43d-4398-a60a-86c1c8101e8a-kube-api-access-l6548\") pod \"nova-cell1-db-create-nnzr4\" (UID: \"34b10dbc-a43d-4398-a60a-86c1c8101e8a\") " pod="openstack/nova-cell1-db-create-nnzr4" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.418324 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffhp2\" (UniqueName: \"kubernetes.io/projected/d61500fa-1a5e-420a-a0d9-990c98754b13-kube-api-access-ffhp2\") pod \"nova-cell0-db-create-xjw8l\" (UID: \"d61500fa-1a5e-420a-a0d9-990c98754b13\") " pod="openstack/nova-cell0-db-create-xjw8l" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.491483 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xjw8l" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.500787 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6548\" (UniqueName: \"kubernetes.io/projected/34b10dbc-a43d-4398-a60a-86c1c8101e8a-kube-api-access-l6548\") pod \"nova-cell1-db-create-nnzr4\" (UID: \"34b10dbc-a43d-4398-a60a-86c1c8101e8a\") " pod="openstack/nova-cell1-db-create-nnzr4" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.521845 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6548\" (UniqueName: \"kubernetes.io/projected/34b10dbc-a43d-4398-a60a-86c1c8101e8a-kube-api-access-l6548\") pod \"nova-cell1-db-create-nnzr4\" (UID: \"34b10dbc-a43d-4398-a60a-86c1c8101e8a\") " pod="openstack/nova-cell1-db-create-nnzr4" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.670971 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nnzr4" Sep 30 08:23:32 crc kubenswrapper[4810]: I0930 08:23:32.900495 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8q22m"] Sep 30 08:23:33 crc kubenswrapper[4810]: I0930 08:23:33.022976 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8q22m" event={"ID":"c40b4429-5a41-4ea8-b9ac-17bba8788f6a","Type":"ContainerStarted","Data":"649ad4702e7d7be8b4568822747e207989d1ad6ee654095688f8d35f22dada32"} Sep 30 08:23:33 crc kubenswrapper[4810]: I0930 08:23:33.051945 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xjw8l"] Sep 30 08:23:33 crc kubenswrapper[4810]: I0930 08:23:33.258872 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-nnzr4"] Sep 30 08:23:33 crc kubenswrapper[4810]: W0930 08:23:33.387354 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34b10dbc_a43d_4398_a60a_86c1c8101e8a.slice/crio-2031bf887fcda8d6f5dae32b9a8cd52e26c1586915c08dea2ec3f37f55619365 WatchSource:0}: Error finding container 2031bf887fcda8d6f5dae32b9a8cd52e26c1586915c08dea2ec3f37f55619365: Status 404 returned error can't find the container with id 2031bf887fcda8d6f5dae32b9a8cd52e26c1586915c08dea2ec3f37f55619365 Sep 30 08:23:34 crc kubenswrapper[4810]: I0930 08:23:34.038318 4810 generic.go:334] "Generic (PLEG): container finished" podID="d61500fa-1a5e-420a-a0d9-990c98754b13" containerID="2dd21650479bf94a3811552d259d60b9c28a18aa4c63d3352ff8e3e1adef3397" exitCode=0 Sep 30 08:23:34 crc kubenswrapper[4810]: I0930 08:23:34.038418 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xjw8l" event={"ID":"d61500fa-1a5e-420a-a0d9-990c98754b13","Type":"ContainerDied","Data":"2dd21650479bf94a3811552d259d60b9c28a18aa4c63d3352ff8e3e1adef3397"} Sep 30 08:23:34 crc kubenswrapper[4810]: I0930 08:23:34.038695 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xjw8l" event={"ID":"d61500fa-1a5e-420a-a0d9-990c98754b13","Type":"ContainerStarted","Data":"5802238b44118d3d97ec202d16df1975faeba7f3127e3884a52561bd1f002f0f"} Sep 30 08:23:34 crc kubenswrapper[4810]: I0930 08:23:34.040939 4810 generic.go:334] "Generic (PLEG): container finished" podID="c40b4429-5a41-4ea8-b9ac-17bba8788f6a" containerID="19ed5ba96885f13d5d5fd5ea34dd10a1a27ae7457de53eff539bdf7679e0114d" exitCode=0 Sep 30 08:23:34 crc kubenswrapper[4810]: I0930 08:23:34.041051 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8q22m" event={"ID":"c40b4429-5a41-4ea8-b9ac-17bba8788f6a","Type":"ContainerDied","Data":"19ed5ba96885f13d5d5fd5ea34dd10a1a27ae7457de53eff539bdf7679e0114d"} Sep 30 08:23:34 crc kubenswrapper[4810]: I0930 08:23:34.043353 4810 generic.go:334] "Generic (PLEG): container finished" podID="34b10dbc-a43d-4398-a60a-86c1c8101e8a" containerID="36de72ecd40625a782355017c702907934cf07fad7f9841c55c526dd9260b553" exitCode=0 Sep 30 08:23:34 crc kubenswrapper[4810]: I0930 08:23:34.043400 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nnzr4" event={"ID":"34b10dbc-a43d-4398-a60a-86c1c8101e8a","Type":"ContainerDied","Data":"36de72ecd40625a782355017c702907934cf07fad7f9841c55c526dd9260b553"} Sep 30 08:23:34 crc kubenswrapper[4810]: I0930 08:23:34.043430 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nnzr4" event={"ID":"34b10dbc-a43d-4398-a60a-86c1c8101e8a","Type":"ContainerStarted","Data":"2031bf887fcda8d6f5dae32b9a8cd52e26c1586915c08dea2ec3f37f55619365"} Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.056974 4810 generic.go:334] "Generic (PLEG): container finished" podID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerID="fa1405eac25bff82b76214a60c2e7583761f1d6bb64e477c50c3991118f3ed4b" exitCode=0 Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.057056 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe2f72a-f3b3-486f-8226-400a43ea161a","Type":"ContainerDied","Data":"fa1405eac25bff82b76214a60c2e7583761f1d6bb64e477c50c3991118f3ed4b"} Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.057651 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbe2f72a-f3b3-486f-8226-400a43ea161a","Type":"ContainerDied","Data":"eee6123203b2df6fc3230b44aaaa1bbea81de0c38ca22b557142c9da14dcfa72"} Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.057680 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eee6123203b2df6fc3230b44aaaa1bbea81de0c38ca22b557142c9da14dcfa72" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.175518 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.262443 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-config-data\") pod \"bbe2f72a-f3b3-486f-8226-400a43ea161a\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.262488 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-combined-ca-bundle\") pod \"bbe2f72a-f3b3-486f-8226-400a43ea161a\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.262636 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-sg-core-conf-yaml\") pod \"bbe2f72a-f3b3-486f-8226-400a43ea161a\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.262662 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-log-httpd\") pod \"bbe2f72a-f3b3-486f-8226-400a43ea161a\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.262770 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-run-httpd\") pod \"bbe2f72a-f3b3-486f-8226-400a43ea161a\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.262801 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dzvm\" (UniqueName: \"kubernetes.io/projected/bbe2f72a-f3b3-486f-8226-400a43ea161a-kube-api-access-4dzvm\") pod \"bbe2f72a-f3b3-486f-8226-400a43ea161a\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.262837 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-scripts\") pod \"bbe2f72a-f3b3-486f-8226-400a43ea161a\" (UID: \"bbe2f72a-f3b3-486f-8226-400a43ea161a\") " Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.281005 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bbe2f72a-f3b3-486f-8226-400a43ea161a" (UID: "bbe2f72a-f3b3-486f-8226-400a43ea161a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.281446 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bbe2f72a-f3b3-486f-8226-400a43ea161a" (UID: "bbe2f72a-f3b3-486f-8226-400a43ea161a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.293890 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-scripts" (OuterVolumeSpecName: "scripts") pod "bbe2f72a-f3b3-486f-8226-400a43ea161a" (UID: "bbe2f72a-f3b3-486f-8226-400a43ea161a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.293918 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbe2f72a-f3b3-486f-8226-400a43ea161a-kube-api-access-4dzvm" (OuterVolumeSpecName: "kube-api-access-4dzvm") pod "bbe2f72a-f3b3-486f-8226-400a43ea161a" (UID: "bbe2f72a-f3b3-486f-8226-400a43ea161a"). InnerVolumeSpecName "kube-api-access-4dzvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.333679 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bbe2f72a-f3b3-486f-8226-400a43ea161a" (UID: "bbe2f72a-f3b3-486f-8226-400a43ea161a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.379838 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.379882 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.379895 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbe2f72a-f3b3-486f-8226-400a43ea161a-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.379907 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dzvm\" (UniqueName: \"kubernetes.io/projected/bbe2f72a-f3b3-486f-8226-400a43ea161a-kube-api-access-4dzvm\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.379921 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.467402 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbe2f72a-f3b3-486f-8226-400a43ea161a" (UID: "bbe2f72a-f3b3-486f-8226-400a43ea161a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.483647 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.538356 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8q22m" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.564218 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-config-data" (OuterVolumeSpecName: "config-data") pod "bbe2f72a-f3b3-486f-8226-400a43ea161a" (UID: "bbe2f72a-f3b3-486f-8226-400a43ea161a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.587141 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbe2f72a-f3b3-486f-8226-400a43ea161a-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.588938 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xjw8l" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.599156 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nnzr4" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.688346 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgp6c\" (UniqueName: \"kubernetes.io/projected/c40b4429-5a41-4ea8-b9ac-17bba8788f6a-kube-api-access-wgp6c\") pod \"c40b4429-5a41-4ea8-b9ac-17bba8788f6a\" (UID: \"c40b4429-5a41-4ea8-b9ac-17bba8788f6a\") " Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.688611 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6548\" (UniqueName: \"kubernetes.io/projected/34b10dbc-a43d-4398-a60a-86c1c8101e8a-kube-api-access-l6548\") pod \"34b10dbc-a43d-4398-a60a-86c1c8101e8a\" (UID: \"34b10dbc-a43d-4398-a60a-86c1c8101e8a\") " Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.688857 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffhp2\" (UniqueName: \"kubernetes.io/projected/d61500fa-1a5e-420a-a0d9-990c98754b13-kube-api-access-ffhp2\") pod \"d61500fa-1a5e-420a-a0d9-990c98754b13\" (UID: \"d61500fa-1a5e-420a-a0d9-990c98754b13\") " Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.693802 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34b10dbc-a43d-4398-a60a-86c1c8101e8a-kube-api-access-l6548" (OuterVolumeSpecName: "kube-api-access-l6548") pod "34b10dbc-a43d-4398-a60a-86c1c8101e8a" (UID: "34b10dbc-a43d-4398-a60a-86c1c8101e8a"). InnerVolumeSpecName "kube-api-access-l6548". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.698445 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c40b4429-5a41-4ea8-b9ac-17bba8788f6a-kube-api-access-wgp6c" (OuterVolumeSpecName: "kube-api-access-wgp6c") pod "c40b4429-5a41-4ea8-b9ac-17bba8788f6a" (UID: "c40b4429-5a41-4ea8-b9ac-17bba8788f6a"). InnerVolumeSpecName "kube-api-access-wgp6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.698497 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d61500fa-1a5e-420a-a0d9-990c98754b13-kube-api-access-ffhp2" (OuterVolumeSpecName: "kube-api-access-ffhp2") pod "d61500fa-1a5e-420a-a0d9-990c98754b13" (UID: "d61500fa-1a5e-420a-a0d9-990c98754b13"). InnerVolumeSpecName "kube-api-access-ffhp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.791048 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgp6c\" (UniqueName: \"kubernetes.io/projected/c40b4429-5a41-4ea8-b9ac-17bba8788f6a-kube-api-access-wgp6c\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.791087 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6548\" (UniqueName: \"kubernetes.io/projected/34b10dbc-a43d-4398-a60a-86c1c8101e8a-kube-api-access-l6548\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:35 crc kubenswrapper[4810]: I0930 08:23:35.791108 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffhp2\" (UniqueName: \"kubernetes.io/projected/d61500fa-1a5e-420a-a0d9-990c98754b13-kube-api-access-ffhp2\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.066925 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xjw8l" event={"ID":"d61500fa-1a5e-420a-a0d9-990c98754b13","Type":"ContainerDied","Data":"5802238b44118d3d97ec202d16df1975faeba7f3127e3884a52561bd1f002f0f"} Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.067005 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5802238b44118d3d97ec202d16df1975faeba7f3127e3884a52561bd1f002f0f" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.066951 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xjw8l" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.068743 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8q22m" event={"ID":"c40b4429-5a41-4ea8-b9ac-17bba8788f6a","Type":"ContainerDied","Data":"649ad4702e7d7be8b4568822747e207989d1ad6ee654095688f8d35f22dada32"} Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.068776 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="649ad4702e7d7be8b4568822747e207989d1ad6ee654095688f8d35f22dada32" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.068826 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8q22m" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.074878 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nnzr4" event={"ID":"34b10dbc-a43d-4398-a60a-86c1c8101e8a","Type":"ContainerDied","Data":"2031bf887fcda8d6f5dae32b9a8cd52e26c1586915c08dea2ec3f37f55619365"} Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.075104 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2031bf887fcda8d6f5dae32b9a8cd52e26c1586915c08dea2ec3f37f55619365" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.074916 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nnzr4" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.074911 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.154594 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.172061 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.188678 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:36 crc kubenswrapper[4810]: E0930 08:23:36.189143 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="proxy-httpd" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189161 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="proxy-httpd" Sep 30 08:23:36 crc kubenswrapper[4810]: E0930 08:23:36.189187 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b10dbc-a43d-4398-a60a-86c1c8101e8a" containerName="mariadb-database-create" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189193 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b10dbc-a43d-4398-a60a-86c1c8101e8a" containerName="mariadb-database-create" Sep 30 08:23:36 crc kubenswrapper[4810]: E0930 08:23:36.189203 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="ceilometer-notification-agent" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189209 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="ceilometer-notification-agent" Sep 30 08:23:36 crc kubenswrapper[4810]: E0930 08:23:36.189228 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d61500fa-1a5e-420a-a0d9-990c98754b13" containerName="mariadb-database-create" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189234 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d61500fa-1a5e-420a-a0d9-990c98754b13" containerName="mariadb-database-create" Sep 30 08:23:36 crc kubenswrapper[4810]: E0930 08:23:36.189249 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="ceilometer-central-agent" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189255 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="ceilometer-central-agent" Sep 30 08:23:36 crc kubenswrapper[4810]: E0930 08:23:36.189281 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="sg-core" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189287 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="sg-core" Sep 30 08:23:36 crc kubenswrapper[4810]: E0930 08:23:36.189295 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c40b4429-5a41-4ea8-b9ac-17bba8788f6a" containerName="mariadb-database-create" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189301 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c40b4429-5a41-4ea8-b9ac-17bba8788f6a" containerName="mariadb-database-create" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189480 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="ceilometer-notification-agent" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189492 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="proxy-httpd" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189500 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="sg-core" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189515 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" containerName="ceilometer-central-agent" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189526 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c40b4429-5a41-4ea8-b9ac-17bba8788f6a" containerName="mariadb-database-create" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189531 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d61500fa-1a5e-420a-a0d9-990c98754b13" containerName="mariadb-database-create" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.189542 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b10dbc-a43d-4398-a60a-86c1c8101e8a" containerName="mariadb-database-create" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.202077 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.206582 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.206757 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.210954 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.289947 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.290000 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.305558 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-log-httpd\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.305604 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.305631 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-run-httpd\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.305882 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-scripts\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.305971 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sglxb\" (UniqueName: \"kubernetes.io/projected/35f89caa-b943-46d8-bfa6-63c88e0a9e02-kube-api-access-sglxb\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.306004 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-config-data\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.306097 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.318689 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.327979 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.408023 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.408117 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-log-httpd\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.408141 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.408699 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-run-httpd\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.408695 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-log-httpd\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.408164 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-run-httpd\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.408959 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-scripts\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.409007 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sglxb\" (UniqueName: \"kubernetes.io/projected/35f89caa-b943-46d8-bfa6-63c88e0a9e02-kube-api-access-sglxb\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.409025 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-config-data\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.412531 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.414542 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-scripts\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.414925 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-config-data\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.416490 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.438525 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sglxb\" (UniqueName: \"kubernetes.io/projected/35f89caa-b943-46d8-bfa6-63c88e0a9e02-kube-api-access-sglxb\") pod \"ceilometer-0\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " pod="openstack/ceilometer-0" Sep 30 08:23:36 crc kubenswrapper[4810]: I0930 08:23:36.520772 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:37 crc kubenswrapper[4810]: I0930 08:23:37.012649 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:37 crc kubenswrapper[4810]: I0930 08:23:37.087615 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35f89caa-b943-46d8-bfa6-63c88e0a9e02","Type":"ContainerStarted","Data":"4a70e0ff7abe54a471c9a89b08c73d50b481677c84a750e99e36802533337aab"} Sep 30 08:23:37 crc kubenswrapper[4810]: I0930 08:23:37.090024 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 08:23:37 crc kubenswrapper[4810]: I0930 08:23:37.090055 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 08:23:37 crc kubenswrapper[4810]: I0930 08:23:37.330344 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbe2f72a-f3b3-486f-8226-400a43ea161a" path="/var/lib/kubelet/pods/bbe2f72a-f3b3-486f-8226-400a43ea161a/volumes" Sep 30 08:23:37 crc kubenswrapper[4810]: I0930 08:23:37.523231 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Sep 30 08:23:38 crc kubenswrapper[4810]: I0930 08:23:38.098142 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35f89caa-b943-46d8-bfa6-63c88e0a9e02","Type":"ContainerStarted","Data":"4b5b33687536854514c60daee6f8030f5418f5d647a7201908c648342095243e"} Sep 30 08:23:38 crc kubenswrapper[4810]: I0930 08:23:38.098526 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35f89caa-b943-46d8-bfa6-63c88e0a9e02","Type":"ContainerStarted","Data":"e1fef1ab8e9f102cbc404235235d289cd69ed9e217f6f918ee42c890ddb5d7a0"} Sep 30 08:23:39 crc kubenswrapper[4810]: I0930 08:23:39.032995 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 08:23:39 crc kubenswrapper[4810]: I0930 08:23:39.039103 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 08:23:39 crc kubenswrapper[4810]: I0930 08:23:39.130653 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35f89caa-b943-46d8-bfa6-63c88e0a9e02","Type":"ContainerStarted","Data":"1daa50c9ea31300d687845e08efedbf534cc24f9b33bd804e2b161486cdf2bd6"} Sep 30 08:23:39 crc kubenswrapper[4810]: I0930 08:23:39.306589 4810 scope.go:117] "RemoveContainer" containerID="ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846" Sep 30 08:23:39 crc kubenswrapper[4810]: E0930 08:23:39.306824 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:23:41 crc kubenswrapper[4810]: I0930 08:23:41.159498 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35f89caa-b943-46d8-bfa6-63c88e0a9e02","Type":"ContainerStarted","Data":"cb287abc98b649811ea4ef9a823c473d2f1257c003ed9d3cb9aabb0815ea7e6b"} Sep 30 08:23:41 crc kubenswrapper[4810]: I0930 08:23:41.160349 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 08:23:41 crc kubenswrapper[4810]: I0930 08:23:41.188322 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.256801697 podStartE2EDuration="5.188299077s" podCreationTimestamp="2025-09-30 08:23:36 +0000 UTC" firstStartedPulling="2025-09-30 08:23:37.018402454 +0000 UTC m=+1240.470601721" lastFinishedPulling="2025-09-30 08:23:39.949899814 +0000 UTC m=+1243.402099101" observedRunningTime="2025-09-30 08:23:41.185140165 +0000 UTC m=+1244.637339432" watchObservedRunningTime="2025-09-30 08:23:41.188299077 +0000 UTC m=+1244.640498344" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.253961 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c663-account-create-hx8d6"] Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.255339 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c663-account-create-hx8d6" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.257797 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.271240 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c663-account-create-hx8d6"] Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.350601 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-694rf\" (UniqueName: \"kubernetes.io/projected/a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb-kube-api-access-694rf\") pod \"nova-api-c663-account-create-hx8d6\" (UID: \"a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb\") " pod="openstack/nova-api-c663-account-create-hx8d6" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.442295 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-00fe-account-create-6v4sg"] Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.444071 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-00fe-account-create-6v4sg" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.450393 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-00fe-account-create-6v4sg"] Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.451065 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.452359 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-694rf\" (UniqueName: \"kubernetes.io/projected/a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb-kube-api-access-694rf\") pod \"nova-api-c663-account-create-hx8d6\" (UID: \"a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb\") " pod="openstack/nova-api-c663-account-create-hx8d6" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.480152 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-694rf\" (UniqueName: \"kubernetes.io/projected/a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb-kube-api-access-694rf\") pod \"nova-api-c663-account-create-hx8d6\" (UID: \"a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb\") " pod="openstack/nova-api-c663-account-create-hx8d6" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.554255 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5t4j\" (UniqueName: \"kubernetes.io/projected/608ee485-d5c0-4724-9101-54958f7344b0-kube-api-access-z5t4j\") pod \"nova-cell0-00fe-account-create-6v4sg\" (UID: \"608ee485-d5c0-4724-9101-54958f7344b0\") " pod="openstack/nova-cell0-00fe-account-create-6v4sg" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.616171 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c663-account-create-hx8d6" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.632887 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-4b13-account-create-t2xtd"] Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.634350 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4b13-account-create-t2xtd" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.642869 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.648961 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4b13-account-create-t2xtd"] Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.658770 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spqv9\" (UniqueName: \"kubernetes.io/projected/913c4569-9c55-4ba3-8a28-b4a23f294852-kube-api-access-spqv9\") pod \"nova-cell1-4b13-account-create-t2xtd\" (UID: \"913c4569-9c55-4ba3-8a28-b4a23f294852\") " pod="openstack/nova-cell1-4b13-account-create-t2xtd" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.659183 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5t4j\" (UniqueName: \"kubernetes.io/projected/608ee485-d5c0-4724-9101-54958f7344b0-kube-api-access-z5t4j\") pod \"nova-cell0-00fe-account-create-6v4sg\" (UID: \"608ee485-d5c0-4724-9101-54958f7344b0\") " pod="openstack/nova-cell0-00fe-account-create-6v4sg" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.689804 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5t4j\" (UniqueName: \"kubernetes.io/projected/608ee485-d5c0-4724-9101-54958f7344b0-kube-api-access-z5t4j\") pod \"nova-cell0-00fe-account-create-6v4sg\" (UID: \"608ee485-d5c0-4724-9101-54958f7344b0\") " pod="openstack/nova-cell0-00fe-account-create-6v4sg" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.761289 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spqv9\" (UniqueName: \"kubernetes.io/projected/913c4569-9c55-4ba3-8a28-b4a23f294852-kube-api-access-spqv9\") pod \"nova-cell1-4b13-account-create-t2xtd\" (UID: \"913c4569-9c55-4ba3-8a28-b4a23f294852\") " pod="openstack/nova-cell1-4b13-account-create-t2xtd" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.786742 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spqv9\" (UniqueName: \"kubernetes.io/projected/913c4569-9c55-4ba3-8a28-b4a23f294852-kube-api-access-spqv9\") pod \"nova-cell1-4b13-account-create-t2xtd\" (UID: \"913c4569-9c55-4ba3-8a28-b4a23f294852\") " pod="openstack/nova-cell1-4b13-account-create-t2xtd" Sep 30 08:23:42 crc kubenswrapper[4810]: I0930 08:23:42.832653 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-00fe-account-create-6v4sg" Sep 30 08:23:43 crc kubenswrapper[4810]: I0930 08:23:43.073718 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4b13-account-create-t2xtd" Sep 30 08:23:43 crc kubenswrapper[4810]: I0930 08:23:43.077526 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c663-account-create-hx8d6"] Sep 30 08:23:43 crc kubenswrapper[4810]: I0930 08:23:43.187729 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c663-account-create-hx8d6" event={"ID":"a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb","Type":"ContainerStarted","Data":"d2786f61d7d2a0aeaa4676d16694c55aac9521b04f788340dfbdbc8bb252d057"} Sep 30 08:23:43 crc kubenswrapper[4810]: I0930 08:23:43.323481 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-00fe-account-create-6v4sg"] Sep 30 08:23:43 crc kubenswrapper[4810]: I0930 08:23:43.478314 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Sep 30 08:23:43 crc kubenswrapper[4810]: I0930 08:23:43.479654 4810 scope.go:117] "RemoveContainer" containerID="ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846" Sep 30 08:23:43 crc kubenswrapper[4810]: E0930 08:23:43.480060 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:23:43 crc kubenswrapper[4810]: I0930 08:23:43.480906 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:23:43 crc kubenswrapper[4810]: I0930 08:23:43.561620 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4b13-account-create-t2xtd"] Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.202041 4810 generic.go:334] "Generic (PLEG): container finished" podID="608ee485-d5c0-4724-9101-54958f7344b0" containerID="8a4ccfcefec56c090ccac05acacdc1358a2cf3204037f22800b3da0ee707741d" exitCode=0 Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.202096 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-00fe-account-create-6v4sg" event={"ID":"608ee485-d5c0-4724-9101-54958f7344b0","Type":"ContainerDied","Data":"8a4ccfcefec56c090ccac05acacdc1358a2cf3204037f22800b3da0ee707741d"} Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.202153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-00fe-account-create-6v4sg" event={"ID":"608ee485-d5c0-4724-9101-54958f7344b0","Type":"ContainerStarted","Data":"a11753243cbafb480efe6dd6ff07b67663e220a24831b5329b3b4a57c3e46ec4"} Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.203983 4810 generic.go:334] "Generic (PLEG): container finished" podID="913c4569-9c55-4ba3-8a28-b4a23f294852" containerID="73e5b89bd1e0fb17af98a0abd0af887cac1e4f3253865d48831c7d567fcaa6c5" exitCode=0 Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.204049 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4b13-account-create-t2xtd" event={"ID":"913c4569-9c55-4ba3-8a28-b4a23f294852","Type":"ContainerDied","Data":"73e5b89bd1e0fb17af98a0abd0af887cac1e4f3253865d48831c7d567fcaa6c5"} Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.204070 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4b13-account-create-t2xtd" event={"ID":"913c4569-9c55-4ba3-8a28-b4a23f294852","Type":"ContainerStarted","Data":"9c5734289a5124a7d6ae7cdfa7a4c64c1f6dde3d365047797e9d1b3e213e902e"} Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.205678 4810 generic.go:334] "Generic (PLEG): container finished" podID="a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb" containerID="17d1f548f5ae13f7f2fb2e80bcc3962ca7cf9e3bb3f204dc17076f9db1acdf15" exitCode=0 Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.206311 4810 scope.go:117] "RemoveContainer" containerID="ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846" Sep 30 08:23:44 crc kubenswrapper[4810]: E0930 08:23:44.206663 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.206941 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c663-account-create-hx8d6" event={"ID":"a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb","Type":"ContainerDied","Data":"17d1f548f5ae13f7f2fb2e80bcc3962ca7cf9e3bb3f204dc17076f9db1acdf15"} Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.824195 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.824545 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="ceilometer-central-agent" containerID="cri-o://e1fef1ab8e9f102cbc404235235d289cd69ed9e217f6f918ee42c890ddb5d7a0" gracePeriod=30 Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.824621 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="proxy-httpd" containerID="cri-o://cb287abc98b649811ea4ef9a823c473d2f1257c003ed9d3cb9aabb0815ea7e6b" gracePeriod=30 Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.824621 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="sg-core" containerID="cri-o://1daa50c9ea31300d687845e08efedbf534cc24f9b33bd804e2b161486cdf2bd6" gracePeriod=30 Sep 30 08:23:44 crc kubenswrapper[4810]: I0930 08:23:44.824621 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="ceilometer-notification-agent" containerID="cri-o://4b5b33687536854514c60daee6f8030f5418f5d647a7201908c648342095243e" gracePeriod=30 Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.217903 4810 generic.go:334] "Generic (PLEG): container finished" podID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerID="cb287abc98b649811ea4ef9a823c473d2f1257c003ed9d3cb9aabb0815ea7e6b" exitCode=0 Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.217963 4810 generic.go:334] "Generic (PLEG): container finished" podID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerID="1daa50c9ea31300d687845e08efedbf534cc24f9b33bd804e2b161486cdf2bd6" exitCode=2 Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.217968 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35f89caa-b943-46d8-bfa6-63c88e0a9e02","Type":"ContainerDied","Data":"cb287abc98b649811ea4ef9a823c473d2f1257c003ed9d3cb9aabb0815ea7e6b"} Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.218003 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35f89caa-b943-46d8-bfa6-63c88e0a9e02","Type":"ContainerDied","Data":"1daa50c9ea31300d687845e08efedbf534cc24f9b33bd804e2b161486cdf2bd6"} Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.702651 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-00fe-account-create-6v4sg" Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.742873 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5t4j\" (UniqueName: \"kubernetes.io/projected/608ee485-d5c0-4724-9101-54958f7344b0-kube-api-access-z5t4j\") pod \"608ee485-d5c0-4724-9101-54958f7344b0\" (UID: \"608ee485-d5c0-4724-9101-54958f7344b0\") " Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.751646 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/608ee485-d5c0-4724-9101-54958f7344b0-kube-api-access-z5t4j" (OuterVolumeSpecName: "kube-api-access-z5t4j") pod "608ee485-d5c0-4724-9101-54958f7344b0" (UID: "608ee485-d5c0-4724-9101-54958f7344b0"). InnerVolumeSpecName "kube-api-access-z5t4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.764356 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4b13-account-create-t2xtd" Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.771241 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c663-account-create-hx8d6" Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.845078 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-694rf\" (UniqueName: \"kubernetes.io/projected/a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb-kube-api-access-694rf\") pod \"a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb\" (UID: \"a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb\") " Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.845530 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spqv9\" (UniqueName: \"kubernetes.io/projected/913c4569-9c55-4ba3-8a28-b4a23f294852-kube-api-access-spqv9\") pod \"913c4569-9c55-4ba3-8a28-b4a23f294852\" (UID: \"913c4569-9c55-4ba3-8a28-b4a23f294852\") " Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.846931 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5t4j\" (UniqueName: \"kubernetes.io/projected/608ee485-d5c0-4724-9101-54958f7344b0-kube-api-access-z5t4j\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.848311 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb-kube-api-access-694rf" (OuterVolumeSpecName: "kube-api-access-694rf") pod "a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb" (UID: "a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb"). InnerVolumeSpecName "kube-api-access-694rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.849577 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/913c4569-9c55-4ba3-8a28-b4a23f294852-kube-api-access-spqv9" (OuterVolumeSpecName: "kube-api-access-spqv9") pod "913c4569-9c55-4ba3-8a28-b4a23f294852" (UID: "913c4569-9c55-4ba3-8a28-b4a23f294852"). InnerVolumeSpecName "kube-api-access-spqv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.911948 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.912010 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.948939 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spqv9\" (UniqueName: \"kubernetes.io/projected/913c4569-9c55-4ba3-8a28-b4a23f294852-kube-api-access-spqv9\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:45 crc kubenswrapper[4810]: I0930 08:23:45.948972 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-694rf\" (UniqueName: \"kubernetes.io/projected/a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb-kube-api-access-694rf\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.234489 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c663-account-create-hx8d6" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.234481 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c663-account-create-hx8d6" event={"ID":"a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb","Type":"ContainerDied","Data":"d2786f61d7d2a0aeaa4676d16694c55aac9521b04f788340dfbdbc8bb252d057"} Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.234644 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2786f61d7d2a0aeaa4676d16694c55aac9521b04f788340dfbdbc8bb252d057" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.243773 4810 generic.go:334] "Generic (PLEG): container finished" podID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerID="4b5b33687536854514c60daee6f8030f5418f5d647a7201908c648342095243e" exitCode=0 Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.243813 4810 generic.go:334] "Generic (PLEG): container finished" podID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerID="e1fef1ab8e9f102cbc404235235d289cd69ed9e217f6f918ee42c890ddb5d7a0" exitCode=0 Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.243855 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35f89caa-b943-46d8-bfa6-63c88e0a9e02","Type":"ContainerDied","Data":"4b5b33687536854514c60daee6f8030f5418f5d647a7201908c648342095243e"} Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.243914 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35f89caa-b943-46d8-bfa6-63c88e0a9e02","Type":"ContainerDied","Data":"e1fef1ab8e9f102cbc404235235d289cd69ed9e217f6f918ee42c890ddb5d7a0"} Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.247082 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-00fe-account-create-6v4sg" event={"ID":"608ee485-d5c0-4724-9101-54958f7344b0","Type":"ContainerDied","Data":"a11753243cbafb480efe6dd6ff07b67663e220a24831b5329b3b4a57c3e46ec4"} Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.247145 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a11753243cbafb480efe6dd6ff07b67663e220a24831b5329b3b4a57c3e46ec4" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.247246 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-00fe-account-create-6v4sg" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.258457 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4b13-account-create-t2xtd" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.258330 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4b13-account-create-t2xtd" event={"ID":"913c4569-9c55-4ba3-8a28-b4a23f294852","Type":"ContainerDied","Data":"9c5734289a5124a7d6ae7cdfa7a4c64c1f6dde3d365047797e9d1b3e213e902e"} Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.260378 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c5734289a5124a7d6ae7cdfa7a4c64c1f6dde3d365047797e9d1b3e213e902e" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.499507 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.567950 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-scripts\") pod \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.568115 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-config-data\") pod \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.568194 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-run-httpd\") pod \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.568316 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-combined-ca-bundle\") pod \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.568346 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-log-httpd\") pod \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.568376 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-sg-core-conf-yaml\") pod \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.568429 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sglxb\" (UniqueName: \"kubernetes.io/projected/35f89caa-b943-46d8-bfa6-63c88e0a9e02-kube-api-access-sglxb\") pod \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\" (UID: \"35f89caa-b943-46d8-bfa6-63c88e0a9e02\") " Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.568917 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "35f89caa-b943-46d8-bfa6-63c88e0a9e02" (UID: "35f89caa-b943-46d8-bfa6-63c88e0a9e02"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.569953 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.571666 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "35f89caa-b943-46d8-bfa6-63c88e0a9e02" (UID: "35f89caa-b943-46d8-bfa6-63c88e0a9e02"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.586447 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35f89caa-b943-46d8-bfa6-63c88e0a9e02-kube-api-access-sglxb" (OuterVolumeSpecName: "kube-api-access-sglxb") pod "35f89caa-b943-46d8-bfa6-63c88e0a9e02" (UID: "35f89caa-b943-46d8-bfa6-63c88e0a9e02"). InnerVolumeSpecName "kube-api-access-sglxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.589414 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-scripts" (OuterVolumeSpecName: "scripts") pod "35f89caa-b943-46d8-bfa6-63c88e0a9e02" (UID: "35f89caa-b943-46d8-bfa6-63c88e0a9e02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.628353 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "35f89caa-b943-46d8-bfa6-63c88e0a9e02" (UID: "35f89caa-b943-46d8-bfa6-63c88e0a9e02"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.671676 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35f89caa-b943-46d8-bfa6-63c88e0a9e02-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.671716 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.671732 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sglxb\" (UniqueName: \"kubernetes.io/projected/35f89caa-b943-46d8-bfa6-63c88e0a9e02-kube-api-access-sglxb\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.671743 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.707560 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35f89caa-b943-46d8-bfa6-63c88e0a9e02" (UID: "35f89caa-b943-46d8-bfa6-63c88e0a9e02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.733030 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-config-data" (OuterVolumeSpecName: "config-data") pod "35f89caa-b943-46d8-bfa6-63c88e0a9e02" (UID: "35f89caa-b943-46d8-bfa6-63c88e0a9e02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.773660 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:46 crc kubenswrapper[4810]: I0930 08:23:46.773689 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35f89caa-b943-46d8-bfa6-63c88e0a9e02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.274430 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35f89caa-b943-46d8-bfa6-63c88e0a9e02","Type":"ContainerDied","Data":"4a70e0ff7abe54a471c9a89b08c73d50b481677c84a750e99e36802533337aab"} Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.274534 4810 scope.go:117] "RemoveContainer" containerID="cb287abc98b649811ea4ef9a823c473d2f1257c003ed9d3cb9aabb0815ea7e6b" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.274829 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.316150 4810 scope.go:117] "RemoveContainer" containerID="1daa50c9ea31300d687845e08efedbf534cc24f9b33bd804e2b161486cdf2bd6" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.328051 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.333423 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.343384 4810 scope.go:117] "RemoveContainer" containerID="4b5b33687536854514c60daee6f8030f5418f5d647a7201908c648342095243e" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.355411 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:47 crc kubenswrapper[4810]: E0930 08:23:47.356473 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913c4569-9c55-4ba3-8a28-b4a23f294852" containerName="mariadb-account-create" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356515 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="913c4569-9c55-4ba3-8a28-b4a23f294852" containerName="mariadb-account-create" Sep 30 08:23:47 crc kubenswrapper[4810]: E0930 08:23:47.356528 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="ceilometer-notification-agent" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356535 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="ceilometer-notification-agent" Sep 30 08:23:47 crc kubenswrapper[4810]: E0930 08:23:47.356550 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="608ee485-d5c0-4724-9101-54958f7344b0" containerName="mariadb-account-create" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356557 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="608ee485-d5c0-4724-9101-54958f7344b0" containerName="mariadb-account-create" Sep 30 08:23:47 crc kubenswrapper[4810]: E0930 08:23:47.356571 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="sg-core" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356577 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="sg-core" Sep 30 08:23:47 crc kubenswrapper[4810]: E0930 08:23:47.356594 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="proxy-httpd" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356600 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="proxy-httpd" Sep 30 08:23:47 crc kubenswrapper[4810]: E0930 08:23:47.356612 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb" containerName="mariadb-account-create" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356618 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb" containerName="mariadb-account-create" Sep 30 08:23:47 crc kubenswrapper[4810]: E0930 08:23:47.356633 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="ceilometer-central-agent" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356638 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="ceilometer-central-agent" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356830 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="ceilometer-central-agent" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356847 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="608ee485-d5c0-4724-9101-54958f7344b0" containerName="mariadb-account-create" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356861 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb" containerName="mariadb-account-create" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356878 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="sg-core" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356886 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="proxy-httpd" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356896 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="913c4569-9c55-4ba3-8a28-b4a23f294852" containerName="mariadb-account-create" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.356903 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" containerName="ceilometer-notification-agent" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.359977 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.362989 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.363507 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.368614 4810 scope.go:117] "RemoveContainer" containerID="e1fef1ab8e9f102cbc404235235d289cd69ed9e217f6f918ee42c890ddb5d7a0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.369630 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.511880 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-config-data\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.512020 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.512052 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-log-httpd\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.512109 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-run-httpd\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.512131 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.512420 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gndj\" (UniqueName: \"kubernetes.io/projected/e3c3d92a-055e-488d-947f-e32908536440-kube-api-access-9gndj\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.512469 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-scripts\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.613778 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gndj\" (UniqueName: \"kubernetes.io/projected/e3c3d92a-055e-488d-947f-e32908536440-kube-api-access-9gndj\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.613844 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-scripts\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.613874 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-config-data\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.613948 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.613975 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-log-httpd\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.614001 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-run-httpd\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.614020 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.615161 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-log-httpd\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.615750 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-run-httpd\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.619953 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-config-data\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.620753 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-scripts\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.620984 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.624478 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.639632 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gndj\" (UniqueName: \"kubernetes.io/projected/e3c3d92a-055e-488d-947f-e32908536440-kube-api-access-9gndj\") pod \"ceilometer-0\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.690438 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mhwt7"] Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.691807 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.693748 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.694131 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-nvbp9" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.695356 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.703288 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mhwt7"] Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.717825 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.717888 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wg77\" (UniqueName: \"kubernetes.io/projected/6cdc8c6a-9e0f-47b0-888d-af03c4339799-kube-api-access-5wg77\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.717949 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-scripts\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.717968 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-config-data\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.744357 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.819410 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.819475 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wg77\" (UniqueName: \"kubernetes.io/projected/6cdc8c6a-9e0f-47b0-888d-af03c4339799-kube-api-access-5wg77\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.819540 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-scripts\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.819565 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-config-data\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.823485 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-scripts\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.824760 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.827536 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-config-data\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:47 crc kubenswrapper[4810]: I0930 08:23:47.840394 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wg77\" (UniqueName: \"kubernetes.io/projected/6cdc8c6a-9e0f-47b0-888d-af03c4339799-kube-api-access-5wg77\") pod \"nova-cell0-conductor-db-sync-mhwt7\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:48 crc kubenswrapper[4810]: I0930 08:23:48.052655 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:23:48 crc kubenswrapper[4810]: I0930 08:23:48.235341 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:48 crc kubenswrapper[4810]: I0930 08:23:48.290969 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3c3d92a-055e-488d-947f-e32908536440","Type":"ContainerStarted","Data":"95e9c7cee81930b50b65d6620b3dd22fb0f7c16e2d9ec058c24036727692c71b"} Sep 30 08:23:48 crc kubenswrapper[4810]: I0930 08:23:48.587562 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mhwt7"] Sep 30 08:23:49 crc kubenswrapper[4810]: I0930 08:23:49.304020 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3c3d92a-055e-488d-947f-e32908536440","Type":"ContainerStarted","Data":"452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171"} Sep 30 08:23:49 crc kubenswrapper[4810]: I0930 08:23:49.304277 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3c3d92a-055e-488d-947f-e32908536440","Type":"ContainerStarted","Data":"0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1"} Sep 30 08:23:49 crc kubenswrapper[4810]: I0930 08:23:49.329589 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35f89caa-b943-46d8-bfa6-63c88e0a9e02" path="/var/lib/kubelet/pods/35f89caa-b943-46d8-bfa6-63c88e0a9e02/volumes" Sep 30 08:23:49 crc kubenswrapper[4810]: I0930 08:23:49.330524 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mhwt7" event={"ID":"6cdc8c6a-9e0f-47b0-888d-af03c4339799","Type":"ContainerStarted","Data":"1b15a57b06a5a1a568180ad09f5ff1fb2532e0f06e3dde46d41593b10b763439"} Sep 30 08:23:50 crc kubenswrapper[4810]: I0930 08:23:50.327678 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3c3d92a-055e-488d-947f-e32908536440","Type":"ContainerStarted","Data":"048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65"} Sep 30 08:23:51 crc kubenswrapper[4810]: I0930 08:23:51.362248 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3c3d92a-055e-488d-947f-e32908536440","Type":"ContainerStarted","Data":"136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5"} Sep 30 08:23:51 crc kubenswrapper[4810]: I0930 08:23:51.362594 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 08:23:51 crc kubenswrapper[4810]: I0930 08:23:51.388554 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.173694348 podStartE2EDuration="4.388532795s" podCreationTimestamp="2025-09-30 08:23:47 +0000 UTC" firstStartedPulling="2025-09-30 08:23:48.253339795 +0000 UTC m=+1251.705539062" lastFinishedPulling="2025-09-30 08:23:50.468178242 +0000 UTC m=+1253.920377509" observedRunningTime="2025-09-30 08:23:51.384168738 +0000 UTC m=+1254.836368005" watchObservedRunningTime="2025-09-30 08:23:51.388532795 +0000 UTC m=+1254.840732062" Sep 30 08:23:51 crc kubenswrapper[4810]: I0930 08:23:51.693962 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:53 crc kubenswrapper[4810]: I0930 08:23:53.380201 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="ceilometer-central-agent" containerID="cri-o://0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1" gracePeriod=30 Sep 30 08:23:53 crc kubenswrapper[4810]: I0930 08:23:53.380219 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="sg-core" containerID="cri-o://048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65" gracePeriod=30 Sep 30 08:23:53 crc kubenswrapper[4810]: I0930 08:23:53.380241 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="ceilometer-notification-agent" containerID="cri-o://452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171" gracePeriod=30 Sep 30 08:23:53 crc kubenswrapper[4810]: I0930 08:23:53.380253 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="proxy-httpd" containerID="cri-o://136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5" gracePeriod=30 Sep 30 08:23:54 crc kubenswrapper[4810]: I0930 08:23:54.397129 4810 generic.go:334] "Generic (PLEG): container finished" podID="e3c3d92a-055e-488d-947f-e32908536440" containerID="136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5" exitCode=0 Sep 30 08:23:54 crc kubenswrapper[4810]: I0930 08:23:54.397500 4810 generic.go:334] "Generic (PLEG): container finished" podID="e3c3d92a-055e-488d-947f-e32908536440" containerID="048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65" exitCode=2 Sep 30 08:23:54 crc kubenswrapper[4810]: I0930 08:23:54.397515 4810 generic.go:334] "Generic (PLEG): container finished" podID="e3c3d92a-055e-488d-947f-e32908536440" containerID="452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171" exitCode=0 Sep 30 08:23:54 crc kubenswrapper[4810]: I0930 08:23:54.397173 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3c3d92a-055e-488d-947f-e32908536440","Type":"ContainerDied","Data":"136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5"} Sep 30 08:23:54 crc kubenswrapper[4810]: I0930 08:23:54.397561 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3c3d92a-055e-488d-947f-e32908536440","Type":"ContainerDied","Data":"048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65"} Sep 30 08:23:54 crc kubenswrapper[4810]: I0930 08:23:54.397575 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3c3d92a-055e-488d-947f-e32908536440","Type":"ContainerDied","Data":"452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171"} Sep 30 08:23:55 crc kubenswrapper[4810]: I0930 08:23:55.306539 4810 scope.go:117] "RemoveContainer" containerID="ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846" Sep 30 08:23:55 crc kubenswrapper[4810]: E0930 08:23:55.306845 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892)\"" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.399344 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.438110 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mhwt7" event={"ID":"6cdc8c6a-9e0f-47b0-888d-af03c4339799","Type":"ContainerStarted","Data":"0e01270242ec97d2e718beaa40e6f09952eddc7a17f4e3492bf1c883b2d0d5ec"} Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.442247 4810 generic.go:334] "Generic (PLEG): container finished" podID="e3c3d92a-055e-488d-947f-e32908536440" containerID="0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1" exitCode=0 Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.442414 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3c3d92a-055e-488d-947f-e32908536440","Type":"ContainerDied","Data":"0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1"} Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.442517 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3c3d92a-055e-488d-947f-e32908536440","Type":"ContainerDied","Data":"95e9c7cee81930b50b65d6620b3dd22fb0f7c16e2d9ec058c24036727692c71b"} Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.442604 4810 scope.go:117] "RemoveContainer" containerID="136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.442802 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.474989 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-combined-ca-bundle\") pod \"e3c3d92a-055e-488d-947f-e32908536440\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.475230 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-log-httpd\") pod \"e3c3d92a-055e-488d-947f-e32908536440\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.475330 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-run-httpd\") pod \"e3c3d92a-055e-488d-947f-e32908536440\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.475515 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gndj\" (UniqueName: \"kubernetes.io/projected/e3c3d92a-055e-488d-947f-e32908536440-kube-api-access-9gndj\") pod \"e3c3d92a-055e-488d-947f-e32908536440\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.475764 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-config-data\") pod \"e3c3d92a-055e-488d-947f-e32908536440\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.475853 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-sg-core-conf-yaml\") pod \"e3c3d92a-055e-488d-947f-e32908536440\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.475957 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-scripts\") pod \"e3c3d92a-055e-488d-947f-e32908536440\" (UID: \"e3c3d92a-055e-488d-947f-e32908536440\") " Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.478749 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e3c3d92a-055e-488d-947f-e32908536440" (UID: "e3c3d92a-055e-488d-947f-e32908536440"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.481494 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e3c3d92a-055e-488d-947f-e32908536440" (UID: "e3c3d92a-055e-488d-947f-e32908536440"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.484423 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-scripts" (OuterVolumeSpecName: "scripts") pod "e3c3d92a-055e-488d-947f-e32908536440" (UID: "e3c3d92a-055e-488d-947f-e32908536440"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.485645 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c3d92a-055e-488d-947f-e32908536440-kube-api-access-9gndj" (OuterVolumeSpecName: "kube-api-access-9gndj") pod "e3c3d92a-055e-488d-947f-e32908536440" (UID: "e3c3d92a-055e-488d-947f-e32908536440"). InnerVolumeSpecName "kube-api-access-9gndj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.495025 4810 scope.go:117] "RemoveContainer" containerID="048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.526060 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e3c3d92a-055e-488d-947f-e32908536440" (UID: "e3c3d92a-055e-488d-947f-e32908536440"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.578931 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gndj\" (UniqueName: \"kubernetes.io/projected/e3c3d92a-055e-488d-947f-e32908536440-kube-api-access-9gndj\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.578972 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.578986 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.579000 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.579012 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3c3d92a-055e-488d-947f-e32908536440-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.583586 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3c3d92a-055e-488d-947f-e32908536440" (UID: "e3c3d92a-055e-488d-947f-e32908536440"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.619643 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-config-data" (OuterVolumeSpecName: "config-data") pod "e3c3d92a-055e-488d-947f-e32908536440" (UID: "e3c3d92a-055e-488d-947f-e32908536440"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.680869 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.680909 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c3d92a-055e-488d-947f-e32908536440-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.685659 4810 scope.go:117] "RemoveContainer" containerID="452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.706466 4810 scope.go:117] "RemoveContainer" containerID="0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.726122 4810 scope.go:117] "RemoveContainer" containerID="136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5" Sep 30 08:23:58 crc kubenswrapper[4810]: E0930 08:23:58.726534 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5\": container with ID starting with 136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5 not found: ID does not exist" containerID="136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.726583 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5"} err="failed to get container status \"136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5\": rpc error: code = NotFound desc = could not find container \"136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5\": container with ID starting with 136c943b54f8f156723ded3dfef85af5fd2152ee32136d0eb71dd103b02c66b5 not found: ID does not exist" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.726610 4810 scope.go:117] "RemoveContainer" containerID="048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65" Sep 30 08:23:58 crc kubenswrapper[4810]: E0930 08:23:58.726816 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65\": container with ID starting with 048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65 not found: ID does not exist" containerID="048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.726895 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65"} err="failed to get container status \"048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65\": rpc error: code = NotFound desc = could not find container \"048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65\": container with ID starting with 048d303f0ee36097d06d0987b54c606806eecc17935af375ed73a1de94381c65 not found: ID does not exist" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.726966 4810 scope.go:117] "RemoveContainer" containerID="452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171" Sep 30 08:23:58 crc kubenswrapper[4810]: E0930 08:23:58.727237 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171\": container with ID starting with 452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171 not found: ID does not exist" containerID="452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.727292 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171"} err="failed to get container status \"452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171\": rpc error: code = NotFound desc = could not find container \"452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171\": container with ID starting with 452ab9242061a8fb8e265138d4f36057c3d25b93c18d231c678e38477bc2a171 not found: ID does not exist" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.727316 4810 scope.go:117] "RemoveContainer" containerID="0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1" Sep 30 08:23:58 crc kubenswrapper[4810]: E0930 08:23:58.727575 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1\": container with ID starting with 0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1 not found: ID does not exist" containerID="0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.727612 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1"} err="failed to get container status \"0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1\": rpc error: code = NotFound desc = could not find container \"0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1\": container with ID starting with 0f766c406285d2c28c22b698126363b68a4f05fca211d87775ad754250114be1 not found: ID does not exist" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.777319 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-mhwt7" podStartSLOduration=2.60294951 podStartE2EDuration="11.777292868s" podCreationTimestamp="2025-09-30 08:23:47 +0000 UTC" firstStartedPulling="2025-09-30 08:23:48.595436865 +0000 UTC m=+1252.047636132" lastFinishedPulling="2025-09-30 08:23:57.769780223 +0000 UTC m=+1261.221979490" observedRunningTime="2025-09-30 08:23:58.477702974 +0000 UTC m=+1261.929902241" watchObservedRunningTime="2025-09-30 08:23:58.777292868 +0000 UTC m=+1262.229492135" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.788213 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.820487 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.828328 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:58 crc kubenswrapper[4810]: E0930 08:23:58.828841 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="proxy-httpd" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.828862 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="proxy-httpd" Sep 30 08:23:58 crc kubenswrapper[4810]: E0930 08:23:58.828899 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="sg-core" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.828908 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="sg-core" Sep 30 08:23:58 crc kubenswrapper[4810]: E0930 08:23:58.828938 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="ceilometer-notification-agent" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.828946 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="ceilometer-notification-agent" Sep 30 08:23:58 crc kubenswrapper[4810]: E0930 08:23:58.828964 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="ceilometer-central-agent" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.828972 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="ceilometer-central-agent" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.829208 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="ceilometer-notification-agent" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.829229 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="sg-core" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.829252 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="ceilometer-central-agent" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.829290 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c3d92a-055e-488d-947f-e32908536440" containerName="proxy-httpd" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.831684 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.836790 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.838125 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.838131 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.891752 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9n99\" (UniqueName: \"kubernetes.io/projected/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-kube-api-access-t9n99\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.891884 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-run-httpd\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.891915 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-log-httpd\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.891959 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-scripts\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.892017 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.892074 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-config-data\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.892095 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.993036 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9n99\" (UniqueName: \"kubernetes.io/projected/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-kube-api-access-t9n99\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.993089 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-run-httpd\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.993107 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-log-httpd\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.993138 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-scripts\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.993178 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.993219 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-config-data\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.993236 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.993611 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-run-httpd\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.993805 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-log-httpd\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.997185 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.998453 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-config-data\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:58 crc kubenswrapper[4810]: I0930 08:23:58.999137 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:59 crc kubenswrapper[4810]: I0930 08:23:59.001089 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-scripts\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:59 crc kubenswrapper[4810]: I0930 08:23:59.008499 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9n99\" (UniqueName: \"kubernetes.io/projected/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-kube-api-access-t9n99\") pod \"ceilometer-0\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " pod="openstack/ceilometer-0" Sep 30 08:23:59 crc kubenswrapper[4810]: I0930 08:23:59.169704 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:23:59 crc kubenswrapper[4810]: I0930 08:23:59.327570 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3c3d92a-055e-488d-947f-e32908536440" path="/var/lib/kubelet/pods/e3c3d92a-055e-488d-947f-e32908536440/volumes" Sep 30 08:23:59 crc kubenswrapper[4810]: I0930 08:23:59.705324 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:23:59 crc kubenswrapper[4810]: W0930 08:23:59.711465 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6372d9e0_5994_4f45_8e6c_d126c9bfbdc3.slice/crio-4c4f299bb218fcbdc5b7b60e0952ecf3a4a0fdac7adee4c90e2587f7618e51f8 WatchSource:0}: Error finding container 4c4f299bb218fcbdc5b7b60e0952ecf3a4a0fdac7adee4c90e2587f7618e51f8: Status 404 returned error can't find the container with id 4c4f299bb218fcbdc5b7b60e0952ecf3a4a0fdac7adee4c90e2587f7618e51f8 Sep 30 08:24:00 crc kubenswrapper[4810]: I0930 08:24:00.465539 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3","Type":"ContainerStarted","Data":"4b0eb2e8c789b768a3b2b5204b4e1f11b0ddae88b366291c94522d464110a616"} Sep 30 08:24:00 crc kubenswrapper[4810]: I0930 08:24:00.465830 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3","Type":"ContainerStarted","Data":"2ea9647b347e72f0da13d4bcc352258e5241434ca899960e013eb65abfb2de42"} Sep 30 08:24:00 crc kubenswrapper[4810]: I0930 08:24:00.465842 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3","Type":"ContainerStarted","Data":"4c4f299bb218fcbdc5b7b60e0952ecf3a4a0fdac7adee4c90e2587f7618e51f8"} Sep 30 08:24:01 crc kubenswrapper[4810]: I0930 08:24:01.484381 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3","Type":"ContainerStarted","Data":"0cdcfe02e4b69f677ff022b9d8c8f704147d4b38d2336beb526e85f05ba447ac"} Sep 30 08:24:02 crc kubenswrapper[4810]: I0930 08:24:02.497154 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3","Type":"ContainerStarted","Data":"4d05bb820cf7cd66583e21de9f009f14c384a982e01de30df61b96d413e2cb89"} Sep 30 08:24:02 crc kubenswrapper[4810]: I0930 08:24:02.498165 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 08:24:02 crc kubenswrapper[4810]: I0930 08:24:02.537752 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.190738327 podStartE2EDuration="4.537728894s" podCreationTimestamp="2025-09-30 08:23:58 +0000 UTC" firstStartedPulling="2025-09-30 08:23:59.713308136 +0000 UTC m=+1263.165507423" lastFinishedPulling="2025-09-30 08:24:02.060298723 +0000 UTC m=+1265.512497990" observedRunningTime="2025-09-30 08:24:02.523376867 +0000 UTC m=+1265.975576144" watchObservedRunningTime="2025-09-30 08:24:02.537728894 +0000 UTC m=+1265.989928161" Sep 30 08:24:09 crc kubenswrapper[4810]: I0930 08:24:09.306896 4810 scope.go:117] "RemoveContainer" containerID="ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846" Sep 30 08:24:09 crc kubenswrapper[4810]: I0930 08:24:09.587774 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerStarted","Data":"5978d931809ac4ce2d5fd8e511370b796f4c1dc65aa700b77eaa0decc93569fd"} Sep 30 08:24:13 crc kubenswrapper[4810]: I0930 08:24:13.478398 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:13 crc kubenswrapper[4810]: I0930 08:24:13.479032 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:13 crc kubenswrapper[4810]: I0930 08:24:13.524738 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:13 crc kubenswrapper[4810]: I0930 08:24:13.629425 4810 generic.go:334] "Generic (PLEG): container finished" podID="6cdc8c6a-9e0f-47b0-888d-af03c4339799" containerID="0e01270242ec97d2e718beaa40e6f09952eddc7a17f4e3492bf1c883b2d0d5ec" exitCode=0 Sep 30 08:24:13 crc kubenswrapper[4810]: I0930 08:24:13.629585 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mhwt7" event={"ID":"6cdc8c6a-9e0f-47b0-888d-af03c4339799","Type":"ContainerDied","Data":"0e01270242ec97d2e718beaa40e6f09952eddc7a17f4e3492bf1c883b2d0d5ec"} Sep 30 08:24:13 crc kubenswrapper[4810]: I0930 08:24:13.686634 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:13 crc kubenswrapper[4810]: I0930 08:24:13.727888 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.087035 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.250977 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-combined-ca-bundle\") pod \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.251198 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-config-data\") pod \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.251231 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-scripts\") pod \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.251338 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wg77\" (UniqueName: \"kubernetes.io/projected/6cdc8c6a-9e0f-47b0-888d-af03c4339799-kube-api-access-5wg77\") pod \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\" (UID: \"6cdc8c6a-9e0f-47b0-888d-af03c4339799\") " Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.262443 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-scripts" (OuterVolumeSpecName: "scripts") pod "6cdc8c6a-9e0f-47b0-888d-af03c4339799" (UID: "6cdc8c6a-9e0f-47b0-888d-af03c4339799"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.262497 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cdc8c6a-9e0f-47b0-888d-af03c4339799-kube-api-access-5wg77" (OuterVolumeSpecName: "kube-api-access-5wg77") pod "6cdc8c6a-9e0f-47b0-888d-af03c4339799" (UID: "6cdc8c6a-9e0f-47b0-888d-af03c4339799"). InnerVolumeSpecName "kube-api-access-5wg77". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.302538 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-config-data" (OuterVolumeSpecName: "config-data") pod "6cdc8c6a-9e0f-47b0-888d-af03c4339799" (UID: "6cdc8c6a-9e0f-47b0-888d-af03c4339799"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.349280 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cdc8c6a-9e0f-47b0-888d-af03c4339799" (UID: "6cdc8c6a-9e0f-47b0-888d-af03c4339799"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.353792 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wg77\" (UniqueName: \"kubernetes.io/projected/6cdc8c6a-9e0f-47b0-888d-af03c4339799-kube-api-access-5wg77\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.353822 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.353835 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.353843 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cdc8c6a-9e0f-47b0-888d-af03c4339799-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.650778 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mhwt7" event={"ID":"6cdc8c6a-9e0f-47b0-888d-af03c4339799","Type":"ContainerDied","Data":"1b15a57b06a5a1a568180ad09f5ff1fb2532e0f06e3dde46d41593b10b763439"} Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.650810 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mhwt7" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.650833 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b15a57b06a5a1a568180ad09f5ff1fb2532e0f06e3dde46d41593b10b763439" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.650865 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" containerID="cri-o://5978d931809ac4ce2d5fd8e511370b796f4c1dc65aa700b77eaa0decc93569fd" gracePeriod=30 Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.751887 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 08:24:15 crc kubenswrapper[4810]: E0930 08:24:15.752396 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cdc8c6a-9e0f-47b0-888d-af03c4339799" containerName="nova-cell0-conductor-db-sync" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.752419 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cdc8c6a-9e0f-47b0-888d-af03c4339799" containerName="nova-cell0-conductor-db-sync" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.752719 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cdc8c6a-9e0f-47b0-888d-af03c4339799" containerName="nova-cell0-conductor-db-sync" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.753585 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.756720 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-nvbp9" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.758798 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.769719 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.861911 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f67f47e-4e54-44b8-9596-c612858accdd-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8f67f47e-4e54-44b8-9596-c612858accdd\") " pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.862004 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcc4w\" (UniqueName: \"kubernetes.io/projected/8f67f47e-4e54-44b8-9596-c612858accdd-kube-api-access-jcc4w\") pod \"nova-cell0-conductor-0\" (UID: \"8f67f47e-4e54-44b8-9596-c612858accdd\") " pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.862114 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f67f47e-4e54-44b8-9596-c612858accdd-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8f67f47e-4e54-44b8-9596-c612858accdd\") " pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.912148 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.912236 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.912367 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.913536 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d77e6c1674e48be39d628cea92dce547143c707ab6bb630e41eb5a3542975559"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.913615 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://d77e6c1674e48be39d628cea92dce547143c707ab6bb630e41eb5a3542975559" gracePeriod=600 Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.965050 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f67f47e-4e54-44b8-9596-c612858accdd-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8f67f47e-4e54-44b8-9596-c612858accdd\") " pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.965205 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcc4w\" (UniqueName: \"kubernetes.io/projected/8f67f47e-4e54-44b8-9596-c612858accdd-kube-api-access-jcc4w\") pod \"nova-cell0-conductor-0\" (UID: \"8f67f47e-4e54-44b8-9596-c612858accdd\") " pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.965456 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f67f47e-4e54-44b8-9596-c612858accdd-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8f67f47e-4e54-44b8-9596-c612858accdd\") " pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.970240 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f67f47e-4e54-44b8-9596-c612858accdd-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8f67f47e-4e54-44b8-9596-c612858accdd\") " pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.971231 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f67f47e-4e54-44b8-9596-c612858accdd-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8f67f47e-4e54-44b8-9596-c612858accdd\") " pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:15 crc kubenswrapper[4810]: I0930 08:24:15.983043 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcc4w\" (UniqueName: \"kubernetes.io/projected/8f67f47e-4e54-44b8-9596-c612858accdd-kube-api-access-jcc4w\") pod \"nova-cell0-conductor-0\" (UID: \"8f67f47e-4e54-44b8-9596-c612858accdd\") " pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:16 crc kubenswrapper[4810]: I0930 08:24:16.110792 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:16 crc kubenswrapper[4810]: I0930 08:24:16.604609 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 08:24:16 crc kubenswrapper[4810]: I0930 08:24:16.684334 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="d77e6c1674e48be39d628cea92dce547143c707ab6bb630e41eb5a3542975559" exitCode=0 Sep 30 08:24:16 crc kubenswrapper[4810]: I0930 08:24:16.684426 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"d77e6c1674e48be39d628cea92dce547143c707ab6bb630e41eb5a3542975559"} Sep 30 08:24:16 crc kubenswrapper[4810]: I0930 08:24:16.684461 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"56832a708861cd3bbdd73ca09a144eec1b9cab6e8fd4ceb0ce5cd9871bf86c10"} Sep 30 08:24:16 crc kubenswrapper[4810]: I0930 08:24:16.684482 4810 scope.go:117] "RemoveContainer" containerID="d1ca8f9a2402180e7c60714ae09e65a733987d89bb171e82c451bfe63cec7eed" Sep 30 08:24:16 crc kubenswrapper[4810]: I0930 08:24:16.687620 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8f67f47e-4e54-44b8-9596-c612858accdd","Type":"ContainerStarted","Data":"fbff126ab5cf9355e4de55430fb77dff66b834556bb2d57bcd6ddd6d38291606"} Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.710133 4810 generic.go:334] "Generic (PLEG): container finished" podID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerID="5978d931809ac4ce2d5fd8e511370b796f4c1dc65aa700b77eaa0decc93569fd" exitCode=0 Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.710312 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerDied","Data":"5978d931809ac4ce2d5fd8e511370b796f4c1dc65aa700b77eaa0decc93569fd"} Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.710814 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892","Type":"ContainerDied","Data":"00a6fd5584e4f390a6774bbf247a0d5d9b5bac7c959fd97b7fce55ccefbd375d"} Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.710826 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00a6fd5584e4f390a6774bbf247a0d5d9b5bac7c959fd97b7fce55ccefbd375d" Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.710844 4810 scope.go:117] "RemoveContainer" containerID="ca4595492744f109682824e51226b080902400eb5c2e764bbab5262dfb270846" Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.717651 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8f67f47e-4e54-44b8-9596-c612858accdd","Type":"ContainerStarted","Data":"ddf47b4be94aebb72d9218cf7ae77bd72f598799108ab905cf271d3a0c1df9d6"} Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.718224 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.740468 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.740445807 podStartE2EDuration="2.740445807s" podCreationTimestamp="2025-09-30 08:24:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:24:17.734804983 +0000 UTC m=+1281.187004260" watchObservedRunningTime="2025-09-30 08:24:17.740445807 +0000 UTC m=+1281.192645094" Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.787845 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.899919 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-custom-prometheus-ca\") pod \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.900099 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-config-data\") pod \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.900150 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-logs\") pod \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.900219 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlkbq\" (UniqueName: \"kubernetes.io/projected/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-kube-api-access-nlkbq\") pod \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.900701 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-logs" (OuterVolumeSpecName: "logs") pod "e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" (UID: "e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.900766 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-combined-ca-bundle\") pod \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\" (UID: \"e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892\") " Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.901727 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.905532 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-kube-api-access-nlkbq" (OuterVolumeSpecName: "kube-api-access-nlkbq") pod "e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" (UID: "e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892"). InnerVolumeSpecName "kube-api-access-nlkbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.932164 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" (UID: "e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.935395 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" (UID: "e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:17 crc kubenswrapper[4810]: I0930 08:24:17.959625 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-config-data" (OuterVolumeSpecName: "config-data") pod "e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" (UID: "e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.003128 4810 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.003159 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.003173 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlkbq\" (UniqueName: \"kubernetes.io/projected/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-kube-api-access-nlkbq\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.003185 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.736988 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.771669 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.779017 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.798360 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Sep 30 08:24:18 crc kubenswrapper[4810]: E0930 08:24:18.798831 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.798856 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: E0930 08:24:18.798882 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.798890 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: E0930 08:24:18.798901 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.798909 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: E0930 08:24:18.798942 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.798949 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.799212 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.799360 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.799430 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.799451 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.800965 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.803555 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.827337 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.922949 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c9ec55c-ddec-4cab-8041-6b5a9c192212-logs\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.923906 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c9ec55c-ddec-4cab-8041-6b5a9c192212-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.924082 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67tw7\" (UniqueName: \"kubernetes.io/projected/5c9ec55c-ddec-4cab-8041-6b5a9c192212-kube-api-access-67tw7\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.924305 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5c9ec55c-ddec-4cab-8041-6b5a9c192212-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:18 crc kubenswrapper[4810]: I0930 08:24:18.924418 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c9ec55c-ddec-4cab-8041-6b5a9c192212-config-data\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.026862 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5c9ec55c-ddec-4cab-8041-6b5a9c192212-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.026979 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c9ec55c-ddec-4cab-8041-6b5a9c192212-config-data\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.027174 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c9ec55c-ddec-4cab-8041-6b5a9c192212-logs\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.027240 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c9ec55c-ddec-4cab-8041-6b5a9c192212-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.027404 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67tw7\" (UniqueName: \"kubernetes.io/projected/5c9ec55c-ddec-4cab-8041-6b5a9c192212-kube-api-access-67tw7\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.032087 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c9ec55c-ddec-4cab-8041-6b5a9c192212-logs\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.034895 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c9ec55c-ddec-4cab-8041-6b5a9c192212-config-data\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.035666 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5c9ec55c-ddec-4cab-8041-6b5a9c192212-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.048644 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c9ec55c-ddec-4cab-8041-6b5a9c192212-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.048879 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67tw7\" (UniqueName: \"kubernetes.io/projected/5c9ec55c-ddec-4cab-8041-6b5a9c192212-kube-api-access-67tw7\") pod \"watcher-decision-engine-0\" (UID: \"5c9ec55c-ddec-4cab-8041-6b5a9c192212\") " pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.119702 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.326513 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" path="/var/lib/kubelet/pods/e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892/volumes" Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.645939 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Sep 30 08:24:19 crc kubenswrapper[4810]: W0930 08:24:19.647951 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c9ec55c_ddec_4cab_8041_6b5a9c192212.slice/crio-8134792a81504d68fa754d84fe9eb599fcde46770481d5b354377963559f01bb WatchSource:0}: Error finding container 8134792a81504d68fa754d84fe9eb599fcde46770481d5b354377963559f01bb: Status 404 returned error can't find the container with id 8134792a81504d68fa754d84fe9eb599fcde46770481d5b354377963559f01bb Sep 30 08:24:19 crc kubenswrapper[4810]: I0930 08:24:19.751030 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"5c9ec55c-ddec-4cab-8041-6b5a9c192212","Type":"ContainerStarted","Data":"8134792a81504d68fa754d84fe9eb599fcde46770481d5b354377963559f01bb"} Sep 30 08:24:20 crc kubenswrapper[4810]: I0930 08:24:20.761205 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"5c9ec55c-ddec-4cab-8041-6b5a9c192212","Type":"ContainerStarted","Data":"15bf01ca46a1905185ed3ccbe5a2d7dda387cff7721da3a2a6e6fed017fb7bc1"} Sep 30 08:24:20 crc kubenswrapper[4810]: I0930 08:24:20.805111 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.805090845 podStartE2EDuration="2.805090845s" podCreationTimestamp="2025-09-30 08:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:24:20.792734246 +0000 UTC m=+1284.244933533" watchObservedRunningTime="2025-09-30 08:24:20.805090845 +0000 UTC m=+1284.257290122" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.145800 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.629765 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-xtgps"] Sep 30 08:24:21 crc kubenswrapper[4810]: E0930 08:24:21.630232 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.630257 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.630514 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6acc6a2-b3e3-4a1a-8e4e-a0e8fff58892" containerName="watcher-decision-engine" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.631254 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.633905 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.634665 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.648503 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xtgps"] Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.795172 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.797769 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.810702 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.825391 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-config-data\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.825492 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.825544 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-scripts\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.825652 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v4zf\" (UniqueName: \"kubernetes.io/projected/376d03ec-c5fa-4f85-860f-2e6e659958ae-kube-api-access-5v4zf\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.917939 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.932763 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.932851 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f529970a-5e33-481e-9b68-4b0519947798-logs\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.932874 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-config-data\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.932913 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.932946 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-scripts\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.932970 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr6l7\" (UniqueName: \"kubernetes.io/projected/f529970a-5e33-481e-9b68-4b0519947798-kube-api-access-tr6l7\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.933110 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v4zf\" (UniqueName: \"kubernetes.io/projected/376d03ec-c5fa-4f85-860f-2e6e659958ae-kube-api-access-5v4zf\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.935824 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-config-data\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.947589 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.951088 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-scripts\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.956043 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-config-data\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.957732 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v4zf\" (UniqueName: \"kubernetes.io/projected/376d03ec-c5fa-4f85-860f-2e6e659958ae-kube-api-access-5v4zf\") pod \"nova-cell0-cell-mapping-xtgps\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:21 crc kubenswrapper[4810]: I0930 08:24:21.965465 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.011917 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.013596 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.025792 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.041706 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.044011 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.044107 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-config-data\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.044282 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17a8b230-cfc5-415e-a480-d78e8bdff63c-logs\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.044349 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.044417 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-config-data\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.044451 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5dq2\" (UniqueName: \"kubernetes.io/projected/17a8b230-cfc5-415e-a480-d78e8bdff63c-kube-api-access-v5dq2\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.044481 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f529970a-5e33-481e-9b68-4b0519947798-logs\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.044633 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr6l7\" (UniqueName: \"kubernetes.io/projected/f529970a-5e33-481e-9b68-4b0519947798-kube-api-access-tr6l7\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.047387 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.047740 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f529970a-5e33-481e-9b68-4b0519947798-logs\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.048678 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.050865 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.055100 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.066895 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.068298 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.074936 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-config-data\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.075672 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.082338 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.082935 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr6l7\" (UniqueName: \"kubernetes.io/projected/f529970a-5e33-481e-9b68-4b0519947798-kube-api-access-tr6l7\") pod \"nova-api-0\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " pod="openstack/nova-api-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.095860 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.134105 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.147607 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17a8b230-cfc5-415e-a480-d78e8bdff63c-logs\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.147667 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzptp\" (UniqueName: \"kubernetes.io/projected/17299686-04ed-4de5-b64c-a504a5783e83-kube-api-access-jzptp\") pod \"nova-cell1-novncproxy-0\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.147696 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.147717 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.147745 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-config-data\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.147808 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5dq2\" (UniqueName: \"kubernetes.io/projected/17a8b230-cfc5-415e-a480-d78e8bdff63c-kube-api-access-v5dq2\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.147905 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-config-data\") pod \"nova-scheduler-0\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.147934 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.147952 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcrbq\" (UniqueName: \"kubernetes.io/projected/fc34893b-bbff-4ea5-8c03-57a392f4a788-kube-api-access-rcrbq\") pod \"nova-scheduler-0\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.147972 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.154934 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17a8b230-cfc5-415e-a480-d78e8bdff63c-logs\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.156808 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-config-data\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.157128 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.172255 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5dq2\" (UniqueName: \"kubernetes.io/projected/17a8b230-cfc5-415e-a480-d78e8bdff63c-kube-api-access-v5dq2\") pod \"nova-metadata-0\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.175817 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-569f96c6f5-dprqh"] Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.177786 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.217814 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-569f96c6f5-dprqh"] Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.227325 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251540 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-config\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251596 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-config-data\") pod \"nova-scheduler-0\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251656 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-swift-storage-0\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251690 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251715 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcrbq\" (UniqueName: \"kubernetes.io/projected/fc34893b-bbff-4ea5-8c03-57a392f4a788-kube-api-access-rcrbq\") pod \"nova-scheduler-0\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251791 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzptp\" (UniqueName: \"kubernetes.io/projected/17299686-04ed-4de5-b64c-a504a5783e83-kube-api-access-jzptp\") pod \"nova-cell1-novncproxy-0\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251810 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251827 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251892 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-nb\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251923 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-svc\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251945 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzhr6\" (UniqueName: \"kubernetes.io/projected/8999f8b3-780a-4d07-a81a-e1aea00275c1-kube-api-access-rzhr6\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.251961 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-sb\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.261846 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.261952 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.268076 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-config-data\") pod \"nova-scheduler-0\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.271650 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.278113 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcrbq\" (UniqueName: \"kubernetes.io/projected/fc34893b-bbff-4ea5-8c03-57a392f4a788-kube-api-access-rcrbq\") pod \"nova-scheduler-0\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.283454 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzptp\" (UniqueName: \"kubernetes.io/projected/17299686-04ed-4de5-b64c-a504a5783e83-kube-api-access-jzptp\") pod \"nova-cell1-novncproxy-0\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.293473 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.365329 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-nb\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.365397 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-svc\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.365429 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzhr6\" (UniqueName: \"kubernetes.io/projected/8999f8b3-780a-4d07-a81a-e1aea00275c1-kube-api-access-rzhr6\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.365445 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-sb\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.365484 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-config\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.365539 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-swift-storage-0\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.366850 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-swift-storage-0\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.370010 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-svc\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.372318 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-nb\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.374710 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-config\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.376045 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-sb\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.392361 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzhr6\" (UniqueName: \"kubernetes.io/projected/8999f8b3-780a-4d07-a81a-e1aea00275c1-kube-api-access-rzhr6\") pod \"dnsmasq-dns-569f96c6f5-dprqh\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.543596 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.582183 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xtgps"] Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.594920 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:22 crc kubenswrapper[4810]: W0930 08:24:22.601379 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod376d03ec_c5fa_4f85_860f_2e6e659958ae.slice/crio-8b73a5c68993bd8a4f6f8cbf3b9562024e1b7adc0d8415718c61fc9626ae0664 WatchSource:0}: Error finding container 8b73a5c68993bd8a4f6f8cbf3b9562024e1b7adc0d8415718c61fc9626ae0664: Status 404 returned error can't find the container with id 8b73a5c68993bd8a4f6f8cbf3b9562024e1b7adc0d8415718c61fc9626ae0664 Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.801479 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xtgps" event={"ID":"376d03ec-c5fa-4f85-860f-2e6e659958ae","Type":"ContainerStarted","Data":"8b73a5c68993bd8a4f6f8cbf3b9562024e1b7adc0d8415718c61fc9626ae0664"} Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.868509 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.939730 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:24:22 crc kubenswrapper[4810]: I0930 08:24:22.948914 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.022900 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7l4sp"] Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.024301 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.027644 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.027850 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.038550 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7l4sp"] Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.087321 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-scripts\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.087413 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-config-data\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.087516 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmqfr\" (UniqueName: \"kubernetes.io/projected/528064bb-fbc6-42fb-b46d-e932ad6971b1-kube-api-access-pmqfr\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.087559 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.189477 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-scripts\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.189798 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-config-data\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.189859 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmqfr\" (UniqueName: \"kubernetes.io/projected/528064bb-fbc6-42fb-b46d-e932ad6971b1-kube-api-access-pmqfr\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.189882 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.195924 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-scripts\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.196041 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.211107 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmqfr\" (UniqueName: \"kubernetes.io/projected/528064bb-fbc6-42fb-b46d-e932ad6971b1-kube-api-access-pmqfr\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.215243 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-config-data\") pod \"nova-cell1-conductor-db-sync-7l4sp\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.343737 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.344631 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:23 crc kubenswrapper[4810]: W0930 08:24:23.354874 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17299686_04ed_4de5_b64c_a504a5783e83.slice/crio-02281b3a1b6b65c763effe07b64d4436559d36d8b28e9f4c058a197d992d38fd WatchSource:0}: Error finding container 02281b3a1b6b65c763effe07b64d4436559d36d8b28e9f4c058a197d992d38fd: Status 404 returned error can't find the container with id 02281b3a1b6b65c763effe07b64d4436559d36d8b28e9f4c058a197d992d38fd Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.429516 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-569f96c6f5-dprqh"] Sep 30 08:24:23 crc kubenswrapper[4810]: W0930 08:24:23.435783 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8999f8b3_780a_4d07_a81a_e1aea00275c1.slice/crio-38426afcb6e95c2e747919123d6314d9f682efebcc3a847064e147c407de1e26 WatchSource:0}: Error finding container 38426afcb6e95c2e747919123d6314d9f682efebcc3a847064e147c407de1e26: Status 404 returned error can't find the container with id 38426afcb6e95c2e747919123d6314d9f682efebcc3a847064e147c407de1e26 Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.813445 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xtgps" event={"ID":"376d03ec-c5fa-4f85-860f-2e6e659958ae","Type":"ContainerStarted","Data":"2068763a38313dfabb6c73703f63fd6736a6e3acea867c3941ae6cd2ee63af16"} Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.816246 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f529970a-5e33-481e-9b68-4b0519947798","Type":"ContainerStarted","Data":"23d1dc977cadc63d31e4457bdb9ace8cdf2d8037911fde87d8bbd0f14d3f26a8"} Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.819835 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"17299686-04ed-4de5-b64c-a504a5783e83","Type":"ContainerStarted","Data":"02281b3a1b6b65c763effe07b64d4436559d36d8b28e9f4c058a197d992d38fd"} Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.821418 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fc34893b-bbff-4ea5-8c03-57a392f4a788","Type":"ContainerStarted","Data":"4398d52bafeb96d005e05ece446279a87632ea329cac84847dd6f503fb902ce5"} Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.829341 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-xtgps" podStartSLOduration=2.82931726 podStartE2EDuration="2.82931726s" podCreationTimestamp="2025-09-30 08:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:24:23.826757295 +0000 UTC m=+1287.278956562" watchObservedRunningTime="2025-09-30 08:24:23.82931726 +0000 UTC m=+1287.281516527" Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.848667 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17a8b230-cfc5-415e-a480-d78e8bdff63c","Type":"ContainerStarted","Data":"ecadbf42074cc5637ec261c40f35c3be17a7ed88bf59219018934722eab39459"} Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.856227 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" event={"ID":"8999f8b3-780a-4d07-a81a-e1aea00275c1","Type":"ContainerStarted","Data":"edec53c10496593782bc108fc5055a96e17b2b6b2cbd9935eafa0fbde695852a"} Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.856299 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" event={"ID":"8999f8b3-780a-4d07-a81a-e1aea00275c1","Type":"ContainerStarted","Data":"38426afcb6e95c2e747919123d6314d9f682efebcc3a847064e147c407de1e26"} Sep 30 08:24:23 crc kubenswrapper[4810]: I0930 08:24:23.939807 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7l4sp"] Sep 30 08:24:24 crc kubenswrapper[4810]: I0930 08:24:24.876894 4810 generic.go:334] "Generic (PLEG): container finished" podID="8999f8b3-780a-4d07-a81a-e1aea00275c1" containerID="edec53c10496593782bc108fc5055a96e17b2b6b2cbd9935eafa0fbde695852a" exitCode=0 Sep 30 08:24:24 crc kubenswrapper[4810]: I0930 08:24:24.876931 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" event={"ID":"8999f8b3-780a-4d07-a81a-e1aea00275c1","Type":"ContainerDied","Data":"edec53c10496593782bc108fc5055a96e17b2b6b2cbd9935eafa0fbde695852a"} Sep 30 08:24:25 crc kubenswrapper[4810]: W0930 08:24:25.121262 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod528064bb_fbc6_42fb_b46d_e932ad6971b1.slice/crio-9c1128315f58b24fb99fb5eb3f8e34975c2c6a9bd924219886c5cc218cfdf910 WatchSource:0}: Error finding container 9c1128315f58b24fb99fb5eb3f8e34975c2c6a9bd924219886c5cc218cfdf910: Status 404 returned error can't find the container with id 9c1128315f58b24fb99fb5eb3f8e34975c2c6a9bd924219886c5cc218cfdf910 Sep 30 08:24:25 crc kubenswrapper[4810]: I0930 08:24:25.295588 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:24:25 crc kubenswrapper[4810]: I0930 08:24:25.336784 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 08:24:25 crc kubenswrapper[4810]: I0930 08:24:25.891305 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7l4sp" event={"ID":"528064bb-fbc6-42fb-b46d-e932ad6971b1","Type":"ContainerStarted","Data":"9c1128315f58b24fb99fb5eb3f8e34975c2c6a9bd924219886c5cc218cfdf910"} Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.902141 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7l4sp" event={"ID":"528064bb-fbc6-42fb-b46d-e932ad6971b1","Type":"ContainerStarted","Data":"684c407e93bfe61796c8e02b7f5d80aaffe4c10d01d333e151b35197a331061b"} Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.904582 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"17299686-04ed-4de5-b64c-a504a5783e83","Type":"ContainerStarted","Data":"a6a3507362cecedd81e8b65e684e774019c2c108eb2c9d6e8f2010b460be5ba0"} Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.904670 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="17299686-04ed-4de5-b64c-a504a5783e83" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://a6a3507362cecedd81e8b65e684e774019c2c108eb2c9d6e8f2010b460be5ba0" gracePeriod=30 Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.912042 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fc34893b-bbff-4ea5-8c03-57a392f4a788","Type":"ContainerStarted","Data":"393f71ec03f46749da16b49dc6f6ef9b4d126c88d6c8c1fdb8db6a0af09313ca"} Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.927215 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-7l4sp" podStartSLOduration=4.927192995 podStartE2EDuration="4.927192995s" podCreationTimestamp="2025-09-30 08:24:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:24:26.91703243 +0000 UTC m=+1290.369231697" watchObservedRunningTime="2025-09-30 08:24:26.927192995 +0000 UTC m=+1290.379392262" Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.929830 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17a8b230-cfc5-415e-a480-d78e8bdff63c","Type":"ContainerStarted","Data":"e0f948ff00373a2e51bfff5be02da66a3df8c5dcf94f95481b9b6b625fa15d65"} Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.929877 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17a8b230-cfc5-415e-a480-d78e8bdff63c","Type":"ContainerStarted","Data":"c3662bc080135b038a8b4747a99cc4bf4db82aa7d38b2bcb37cf6066588622b4"} Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.930224 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="17a8b230-cfc5-415e-a480-d78e8bdff63c" containerName="nova-metadata-log" containerID="cri-o://c3662bc080135b038a8b4747a99cc4bf4db82aa7d38b2bcb37cf6066588622b4" gracePeriod=30 Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.930332 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="17a8b230-cfc5-415e-a480-d78e8bdff63c" containerName="nova-metadata-metadata" containerID="cri-o://e0f948ff00373a2e51bfff5be02da66a3df8c5dcf94f95481b9b6b625fa15d65" gracePeriod=30 Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.943331 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" event={"ID":"8999f8b3-780a-4d07-a81a-e1aea00275c1","Type":"ContainerStarted","Data":"8e139db0f1b5cfe766e8092b81405fed58e00152c429a7e5c26b2741746e4fd6"} Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.943481 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.949046 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f529970a-5e33-481e-9b68-4b0519947798","Type":"ContainerStarted","Data":"4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca"} Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.949099 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f529970a-5e33-481e-9b68-4b0519947798","Type":"ContainerStarted","Data":"a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc"} Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.956418 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.862690201 podStartE2EDuration="5.956397624s" podCreationTimestamp="2025-09-30 08:24:21 +0000 UTC" firstStartedPulling="2025-09-30 08:24:22.982615368 +0000 UTC m=+1286.434814635" lastFinishedPulling="2025-09-30 08:24:26.076322791 +0000 UTC m=+1289.528522058" observedRunningTime="2025-09-30 08:24:26.935575389 +0000 UTC m=+1290.387774656" watchObservedRunningTime="2025-09-30 08:24:26.956397624 +0000 UTC m=+1290.408596881" Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.987142 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.293392206 podStartE2EDuration="5.987121587s" podCreationTimestamp="2025-09-30 08:24:21 +0000 UTC" firstStartedPulling="2025-09-30 08:24:23.363478505 +0000 UTC m=+1286.815677772" lastFinishedPulling="2025-09-30 08:24:26.057207886 +0000 UTC m=+1289.509407153" observedRunningTime="2025-09-30 08:24:26.950239175 +0000 UTC m=+1290.402438442" watchObservedRunningTime="2025-09-30 08:24:26.987121587 +0000 UTC m=+1290.439320844" Sep 30 08:24:26 crc kubenswrapper[4810]: I0930 08:24:26.989128 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.864592887 podStartE2EDuration="5.989114815s" podCreationTimestamp="2025-09-30 08:24:21 +0000 UTC" firstStartedPulling="2025-09-30 08:24:22.956889491 +0000 UTC m=+1286.409088758" lastFinishedPulling="2025-09-30 08:24:26.081411419 +0000 UTC m=+1289.533610686" observedRunningTime="2025-09-30 08:24:26.963440978 +0000 UTC m=+1290.415640245" watchObservedRunningTime="2025-09-30 08:24:26.989114815 +0000 UTC m=+1290.441314082" Sep 30 08:24:27 crc kubenswrapper[4810]: I0930 08:24:27.024203 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.866796589 podStartE2EDuration="6.024180363s" podCreationTimestamp="2025-09-30 08:24:21 +0000 UTC" firstStartedPulling="2025-09-30 08:24:22.899856483 +0000 UTC m=+1286.352055740" lastFinishedPulling="2025-09-30 08:24:26.057240247 +0000 UTC m=+1289.509439514" observedRunningTime="2025-09-30 08:24:26.985568071 +0000 UTC m=+1290.437767358" watchObservedRunningTime="2025-09-30 08:24:27.024180363 +0000 UTC m=+1290.476379630" Sep 30 08:24:27 crc kubenswrapper[4810]: I0930 08:24:27.033067 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" podStartSLOduration=6.033047381 podStartE2EDuration="6.033047381s" podCreationTimestamp="2025-09-30 08:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:24:27.009461586 +0000 UTC m=+1290.461660873" watchObservedRunningTime="2025-09-30 08:24:27.033047381 +0000 UTC m=+1290.485246658" Sep 30 08:24:27 crc kubenswrapper[4810]: I0930 08:24:27.228242 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 08:24:27 crc kubenswrapper[4810]: I0930 08:24:27.228608 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 08:24:27 crc kubenswrapper[4810]: I0930 08:24:27.294227 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 08:24:27 crc kubenswrapper[4810]: I0930 08:24:27.544473 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:27 crc kubenswrapper[4810]: I0930 08:24:27.965354 4810 generic.go:334] "Generic (PLEG): container finished" podID="17a8b230-cfc5-415e-a480-d78e8bdff63c" containerID="c3662bc080135b038a8b4747a99cc4bf4db82aa7d38b2bcb37cf6066588622b4" exitCode=143 Sep 30 08:24:27 crc kubenswrapper[4810]: I0930 08:24:27.966843 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17a8b230-cfc5-415e-a480-d78e8bdff63c","Type":"ContainerDied","Data":"c3662bc080135b038a8b4747a99cc4bf4db82aa7d38b2bcb37cf6066588622b4"} Sep 30 08:24:29 crc kubenswrapper[4810]: I0930 08:24:29.120714 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:29 crc kubenswrapper[4810]: I0930 08:24:29.156394 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:29 crc kubenswrapper[4810]: I0930 08:24:29.188810 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 30 08:24:29 crc kubenswrapper[4810]: I0930 08:24:29.985624 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:30 crc kubenswrapper[4810]: I0930 08:24:30.019209 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Sep 30 08:24:32 crc kubenswrapper[4810]: I0930 08:24:32.009167 4810 generic.go:334] "Generic (PLEG): container finished" podID="376d03ec-c5fa-4f85-860f-2e6e659958ae" containerID="2068763a38313dfabb6c73703f63fd6736a6e3acea867c3941ae6cd2ee63af16" exitCode=0 Sep 30 08:24:32 crc kubenswrapper[4810]: I0930 08:24:32.010231 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xtgps" event={"ID":"376d03ec-c5fa-4f85-860f-2e6e659958ae","Type":"ContainerDied","Data":"2068763a38313dfabb6c73703f63fd6736a6e3acea867c3941ae6cd2ee63af16"} Sep 30 08:24:32 crc kubenswrapper[4810]: I0930 08:24:32.135811 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 08:24:32 crc kubenswrapper[4810]: I0930 08:24:32.136149 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 08:24:32 crc kubenswrapper[4810]: I0930 08:24:32.294632 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 08:24:32 crc kubenswrapper[4810]: I0930 08:24:32.328781 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 08:24:32 crc kubenswrapper[4810]: I0930 08:24:32.597195 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:24:32 crc kubenswrapper[4810]: I0930 08:24:32.658871 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5595b896c9-xltdz"] Sep 30 08:24:32 crc kubenswrapper[4810]: I0930 08:24:32.659120 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" podUID="e070c75e-313a-423c-88d7-b4975a8d113d" containerName="dnsmasq-dns" containerID="cri-o://1adc7727d0e20d062781ac5384801b5627a66f65be588683002008197f7425dd" gracePeriod=10 Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.043820 4810 generic.go:334] "Generic (PLEG): container finished" podID="e070c75e-313a-423c-88d7-b4975a8d113d" containerID="1adc7727d0e20d062781ac5384801b5627a66f65be588683002008197f7425dd" exitCode=0 Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.043924 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" event={"ID":"e070c75e-313a-423c-88d7-b4975a8d113d","Type":"ContainerDied","Data":"1adc7727d0e20d062781ac5384801b5627a66f65be588683002008197f7425dd"} Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.111674 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.217605 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f529970a-5e33-481e-9b68-4b0519947798" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.209:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.217939 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f529970a-5e33-481e-9b68-4b0519947798" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.209:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.293059 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.417974 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thk8p\" (UniqueName: \"kubernetes.io/projected/e070c75e-313a-423c-88d7-b4975a8d113d-kube-api-access-thk8p\") pod \"e070c75e-313a-423c-88d7-b4975a8d113d\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.418062 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-svc\") pod \"e070c75e-313a-423c-88d7-b4975a8d113d\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.418125 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-sb\") pod \"e070c75e-313a-423c-88d7-b4975a8d113d\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.418143 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-swift-storage-0\") pod \"e070c75e-313a-423c-88d7-b4975a8d113d\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.418212 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-nb\") pod \"e070c75e-313a-423c-88d7-b4975a8d113d\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.418226 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-config\") pod \"e070c75e-313a-423c-88d7-b4975a8d113d\" (UID: \"e070c75e-313a-423c-88d7-b4975a8d113d\") " Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.433025 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e070c75e-313a-423c-88d7-b4975a8d113d-kube-api-access-thk8p" (OuterVolumeSpecName: "kube-api-access-thk8p") pod "e070c75e-313a-423c-88d7-b4975a8d113d" (UID: "e070c75e-313a-423c-88d7-b4975a8d113d"). InnerVolumeSpecName "kube-api-access-thk8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.505016 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-config" (OuterVolumeSpecName: "config") pod "e070c75e-313a-423c-88d7-b4975a8d113d" (UID: "e070c75e-313a-423c-88d7-b4975a8d113d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.509780 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e070c75e-313a-423c-88d7-b4975a8d113d" (UID: "e070c75e-313a-423c-88d7-b4975a8d113d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.521783 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thk8p\" (UniqueName: \"kubernetes.io/projected/e070c75e-313a-423c-88d7-b4975a8d113d-kube-api-access-thk8p\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.521809 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.521818 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.526863 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e070c75e-313a-423c-88d7-b4975a8d113d" (UID: "e070c75e-313a-423c-88d7-b4975a8d113d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.558081 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e070c75e-313a-423c-88d7-b4975a8d113d" (UID: "e070c75e-313a-423c-88d7-b4975a8d113d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.559878 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.585297 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e070c75e-313a-423c-88d7-b4975a8d113d" (UID: "e070c75e-313a-423c-88d7-b4975a8d113d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.622954 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-combined-ca-bundle\") pod \"376d03ec-c5fa-4f85-860f-2e6e659958ae\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.623154 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-scripts\") pod \"376d03ec-c5fa-4f85-860f-2e6e659958ae\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.623249 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5v4zf\" (UniqueName: \"kubernetes.io/projected/376d03ec-c5fa-4f85-860f-2e6e659958ae-kube-api-access-5v4zf\") pod \"376d03ec-c5fa-4f85-860f-2e6e659958ae\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.623289 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-config-data\") pod \"376d03ec-c5fa-4f85-860f-2e6e659958ae\" (UID: \"376d03ec-c5fa-4f85-860f-2e6e659958ae\") " Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.623689 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.623704 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.623714 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e070c75e-313a-423c-88d7-b4975a8d113d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.629326 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/376d03ec-c5fa-4f85-860f-2e6e659958ae-kube-api-access-5v4zf" (OuterVolumeSpecName: "kube-api-access-5v4zf") pod "376d03ec-c5fa-4f85-860f-2e6e659958ae" (UID: "376d03ec-c5fa-4f85-860f-2e6e659958ae"). InnerVolumeSpecName "kube-api-access-5v4zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.634932 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-scripts" (OuterVolumeSpecName: "scripts") pod "376d03ec-c5fa-4f85-860f-2e6e659958ae" (UID: "376d03ec-c5fa-4f85-860f-2e6e659958ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.636343 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.636666 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="1218e0c8-f487-48d3-ae27-2e7b3aa4d07e" containerName="kube-state-metrics" containerID="cri-o://dcb15d23ac7107388574027e20fb65adbf18bcb2018884e5a58b6b6967ddcfd1" gracePeriod=30 Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.653449 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "376d03ec-c5fa-4f85-860f-2e6e659958ae" (UID: "376d03ec-c5fa-4f85-860f-2e6e659958ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.664497 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-config-data" (OuterVolumeSpecName: "config-data") pod "376d03ec-c5fa-4f85-860f-2e6e659958ae" (UID: "376d03ec-c5fa-4f85-860f-2e6e659958ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.725428 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5v4zf\" (UniqueName: \"kubernetes.io/projected/376d03ec-c5fa-4f85-860f-2e6e659958ae-kube-api-access-5v4zf\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.725466 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.725476 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:33 crc kubenswrapper[4810]: I0930 08:24:33.725487 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/376d03ec-c5fa-4f85-860f-2e6e659958ae-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.058124 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" event={"ID":"e070c75e-313a-423c-88d7-b4975a8d113d","Type":"ContainerDied","Data":"3f0eb234189f0878328e8e05900fb638f809f20471713eb6b79150ff96ddebf1"} Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.058201 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.059257 4810 scope.go:117] "RemoveContainer" containerID="1adc7727d0e20d062781ac5384801b5627a66f65be588683002008197f7425dd" Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.075011 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xtgps" Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.075288 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xtgps" event={"ID":"376d03ec-c5fa-4f85-860f-2e6e659958ae","Type":"ContainerDied","Data":"8b73a5c68993bd8a4f6f8cbf3b9562024e1b7adc0d8415718c61fc9626ae0664"} Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.075334 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b73a5c68993bd8a4f6f8cbf3b9562024e1b7adc0d8415718c61fc9626ae0664" Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.081234 4810 generic.go:334] "Generic (PLEG): container finished" podID="1218e0c8-f487-48d3-ae27-2e7b3aa4d07e" containerID="dcb15d23ac7107388574027e20fb65adbf18bcb2018884e5a58b6b6967ddcfd1" exitCode=2 Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.081370 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1218e0c8-f487-48d3-ae27-2e7b3aa4d07e","Type":"ContainerDied","Data":"dcb15d23ac7107388574027e20fb65adbf18bcb2018884e5a58b6b6967ddcfd1"} Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.216878 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.227078 4810 scope.go:117] "RemoveContainer" containerID="febc3f449b977933be184110ef07d9be808e55cd56857f2b3e523f9c4ece6a32" Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.235448 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5595b896c9-xltdz"] Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.255362 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5595b896c9-xltdz"] Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.262603 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.262987 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f529970a-5e33-481e-9b68-4b0519947798" containerName="nova-api-log" containerID="cri-o://a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc" gracePeriod=30 Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.263088 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f529970a-5e33-481e-9b68-4b0519947798" containerName="nova-api-api" containerID="cri-o://4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca" gracePeriod=30 Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.313147 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.339976 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fspl9\" (UniqueName: \"kubernetes.io/projected/1218e0c8-f487-48d3-ae27-2e7b3aa4d07e-kube-api-access-fspl9\") pod \"1218e0c8-f487-48d3-ae27-2e7b3aa4d07e\" (UID: \"1218e0c8-f487-48d3-ae27-2e7b3aa4d07e\") " Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.353721 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1218e0c8-f487-48d3-ae27-2e7b3aa4d07e-kube-api-access-fspl9" (OuterVolumeSpecName: "kube-api-access-fspl9") pod "1218e0c8-f487-48d3-ae27-2e7b3aa4d07e" (UID: "1218e0c8-f487-48d3-ae27-2e7b3aa4d07e"). InnerVolumeSpecName "kube-api-access-fspl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:24:34 crc kubenswrapper[4810]: I0930 08:24:34.443416 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fspl9\" (UniqueName: \"kubernetes.io/projected/1218e0c8-f487-48d3-ae27-2e7b3aa4d07e-kube-api-access-fspl9\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.098295 4810 generic.go:334] "Generic (PLEG): container finished" podID="f529970a-5e33-481e-9b68-4b0519947798" containerID="a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc" exitCode=143 Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.098428 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f529970a-5e33-481e-9b68-4b0519947798","Type":"ContainerDied","Data":"a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc"} Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.102986 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1218e0c8-f487-48d3-ae27-2e7b3aa4d07e","Type":"ContainerDied","Data":"584a5c9639e46b31a8b121672e99c6a46c246d0a081979d0317e33b3e042b9a7"} Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.103071 4810 scope.go:117] "RemoveContainer" containerID="dcb15d23ac7107388574027e20fb65adbf18bcb2018884e5a58b6b6967ddcfd1" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.102997 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.103310 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="fc34893b-bbff-4ea5-8c03-57a392f4a788" containerName="nova-scheduler-scheduler" containerID="cri-o://393f71ec03f46749da16b49dc6f6ef9b4d126c88d6c8c1fdb8db6a0af09313ca" gracePeriod=30 Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.188613 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.200662 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.222999 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 08:24:35 crc kubenswrapper[4810]: E0930 08:24:35.223914 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e070c75e-313a-423c-88d7-b4975a8d113d" containerName="dnsmasq-dns" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.223959 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e070c75e-313a-423c-88d7-b4975a8d113d" containerName="dnsmasq-dns" Sep 30 08:24:35 crc kubenswrapper[4810]: E0930 08:24:35.223978 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376d03ec-c5fa-4f85-860f-2e6e659958ae" containerName="nova-manage" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.223988 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="376d03ec-c5fa-4f85-860f-2e6e659958ae" containerName="nova-manage" Sep 30 08:24:35 crc kubenswrapper[4810]: E0930 08:24:35.224037 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e070c75e-313a-423c-88d7-b4975a8d113d" containerName="init" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.224046 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e070c75e-313a-423c-88d7-b4975a8d113d" containerName="init" Sep 30 08:24:35 crc kubenswrapper[4810]: E0930 08:24:35.224099 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1218e0c8-f487-48d3-ae27-2e7b3aa4d07e" containerName="kube-state-metrics" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.224109 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1218e0c8-f487-48d3-ae27-2e7b3aa4d07e" containerName="kube-state-metrics" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.224579 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="376d03ec-c5fa-4f85-860f-2e6e659958ae" containerName="nova-manage" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.224621 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e070c75e-313a-423c-88d7-b4975a8d113d" containerName="dnsmasq-dns" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.224642 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1218e0c8-f487-48d3-ae27-2e7b3aa4d07e" containerName="kube-state-metrics" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.225673 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.229827 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.230219 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.233463 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.320031 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1218e0c8-f487-48d3-ae27-2e7b3aa4d07e" path="/var/lib/kubelet/pods/1218e0c8-f487-48d3-ae27-2e7b3aa4d07e/volumes" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.322498 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e070c75e-313a-423c-88d7-b4975a8d113d" path="/var/lib/kubelet/pods/e070c75e-313a-423c-88d7-b4975a8d113d/volumes" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.363245 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0778b82-d141-47f3-b766-a3df63f09d12-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.363600 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0778b82-d141-47f3-b766-a3df63f09d12-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.363694 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hc7b\" (UniqueName: \"kubernetes.io/projected/b0778b82-d141-47f3-b766-a3df63f09d12-kube-api-access-9hc7b\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.363769 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b0778b82-d141-47f3-b766-a3df63f09d12-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.464945 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0778b82-d141-47f3-b766-a3df63f09d12-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.465006 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hc7b\" (UniqueName: \"kubernetes.io/projected/b0778b82-d141-47f3-b766-a3df63f09d12-kube-api-access-9hc7b\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.465037 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b0778b82-d141-47f3-b766-a3df63f09d12-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.465189 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0778b82-d141-47f3-b766-a3df63f09d12-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.471074 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0778b82-d141-47f3-b766-a3df63f09d12-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.471609 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b0778b82-d141-47f3-b766-a3df63f09d12-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.471925 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0778b82-d141-47f3-b766-a3df63f09d12-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.485208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hc7b\" (UniqueName: \"kubernetes.io/projected/b0778b82-d141-47f3-b766-a3df63f09d12-kube-api-access-9hc7b\") pod \"kube-state-metrics-0\" (UID: \"b0778b82-d141-47f3-b766-a3df63f09d12\") " pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.553604 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.979600 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.980166 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="sg-core" containerID="cri-o://0cdcfe02e4b69f677ff022b9d8c8f704147d4b38d2336beb526e85f05ba447ac" gracePeriod=30 Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.980305 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="proxy-httpd" containerID="cri-o://4d05bb820cf7cd66583e21de9f009f14c384a982e01de30df61b96d413e2cb89" gracePeriod=30 Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.980247 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="ceilometer-notification-agent" containerID="cri-o://4b0eb2e8c789b768a3b2b5204b4e1f11b0ddae88b366291c94522d464110a616" gracePeriod=30 Sep 30 08:24:35 crc kubenswrapper[4810]: I0930 08:24:35.980138 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="ceilometer-central-agent" containerID="cri-o://2ea9647b347e72f0da13d4bcc352258e5241434ca899960e013eb65abfb2de42" gracePeriod=30 Sep 30 08:24:36 crc kubenswrapper[4810]: I0930 08:24:36.072353 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 08:24:36 crc kubenswrapper[4810]: W0930 08:24:36.076359 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0778b82_d141_47f3_b766_a3df63f09d12.slice/crio-4669b6cfb0ac1cf2f40089875bb87174c5c8b0e5ce56157038b20bdfc6762dcf WatchSource:0}: Error finding container 4669b6cfb0ac1cf2f40089875bb87174c5c8b0e5ce56157038b20bdfc6762dcf: Status 404 returned error can't find the container with id 4669b6cfb0ac1cf2f40089875bb87174c5c8b0e5ce56157038b20bdfc6762dcf Sep 30 08:24:36 crc kubenswrapper[4810]: I0930 08:24:36.122376 4810 generic.go:334] "Generic (PLEG): container finished" podID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerID="4d05bb820cf7cd66583e21de9f009f14c384a982e01de30df61b96d413e2cb89" exitCode=0 Sep 30 08:24:36 crc kubenswrapper[4810]: I0930 08:24:36.122417 4810 generic.go:334] "Generic (PLEG): container finished" podID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerID="0cdcfe02e4b69f677ff022b9d8c8f704147d4b38d2336beb526e85f05ba447ac" exitCode=2 Sep 30 08:24:36 crc kubenswrapper[4810]: I0930 08:24:36.122452 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3","Type":"ContainerDied","Data":"4d05bb820cf7cd66583e21de9f009f14c384a982e01de30df61b96d413e2cb89"} Sep 30 08:24:36 crc kubenswrapper[4810]: I0930 08:24:36.122493 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3","Type":"ContainerDied","Data":"0cdcfe02e4b69f677ff022b9d8c8f704147d4b38d2336beb526e85f05ba447ac"} Sep 30 08:24:36 crc kubenswrapper[4810]: I0930 08:24:36.127301 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b0778b82-d141-47f3-b766-a3df63f09d12","Type":"ContainerStarted","Data":"4669b6cfb0ac1cf2f40089875bb87174c5c8b0e5ce56157038b20bdfc6762dcf"} Sep 30 08:24:37 crc kubenswrapper[4810]: I0930 08:24:37.148896 4810 generic.go:334] "Generic (PLEG): container finished" podID="528064bb-fbc6-42fb-b46d-e932ad6971b1" containerID="684c407e93bfe61796c8e02b7f5d80aaffe4c10d01d333e151b35197a331061b" exitCode=0 Sep 30 08:24:37 crc kubenswrapper[4810]: I0930 08:24:37.149052 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7l4sp" event={"ID":"528064bb-fbc6-42fb-b46d-e932ad6971b1","Type":"ContainerDied","Data":"684c407e93bfe61796c8e02b7f5d80aaffe4c10d01d333e151b35197a331061b"} Sep 30 08:24:37 crc kubenswrapper[4810]: I0930 08:24:37.156171 4810 generic.go:334] "Generic (PLEG): container finished" podID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerID="2ea9647b347e72f0da13d4bcc352258e5241434ca899960e013eb65abfb2de42" exitCode=0 Sep 30 08:24:37 crc kubenswrapper[4810]: I0930 08:24:37.156329 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3","Type":"ContainerDied","Data":"2ea9647b347e72f0da13d4bcc352258e5241434ca899960e013eb65abfb2de42"} Sep 30 08:24:37 crc kubenswrapper[4810]: E0930 08:24:37.298045 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="393f71ec03f46749da16b49dc6f6ef9b4d126c88d6c8c1fdb8db6a0af09313ca" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 08:24:37 crc kubenswrapper[4810]: E0930 08:24:37.299112 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="393f71ec03f46749da16b49dc6f6ef9b4d126c88d6c8c1fdb8db6a0af09313ca" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 08:24:37 crc kubenswrapper[4810]: E0930 08:24:37.300465 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="393f71ec03f46749da16b49dc6f6ef9b4d126c88d6c8c1fdb8db6a0af09313ca" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 08:24:37 crc kubenswrapper[4810]: E0930 08:24:37.300522 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="fc34893b-bbff-4ea5-8c03-57a392f4a788" containerName="nova-scheduler-scheduler" Sep 30 08:24:37 crc kubenswrapper[4810]: I0930 08:24:37.941237 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:24:37 crc kubenswrapper[4810]: I0930 08:24:37.991506 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5595b896c9-xltdz" podUID="e070c75e-313a-423c-88d7-b4975a8d113d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.182:5353: i/o timeout" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.121443 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f529970a-5e33-481e-9b68-4b0519947798-logs\") pod \"f529970a-5e33-481e-9b68-4b0519947798\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.121655 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-config-data\") pod \"f529970a-5e33-481e-9b68-4b0519947798\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.121697 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr6l7\" (UniqueName: \"kubernetes.io/projected/f529970a-5e33-481e-9b68-4b0519947798-kube-api-access-tr6l7\") pod \"f529970a-5e33-481e-9b68-4b0519947798\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.121739 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-combined-ca-bundle\") pod \"f529970a-5e33-481e-9b68-4b0519947798\" (UID: \"f529970a-5e33-481e-9b68-4b0519947798\") " Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.122766 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f529970a-5e33-481e-9b68-4b0519947798-logs" (OuterVolumeSpecName: "logs") pod "f529970a-5e33-481e-9b68-4b0519947798" (UID: "f529970a-5e33-481e-9b68-4b0519947798"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.129063 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f529970a-5e33-481e-9b68-4b0519947798-kube-api-access-tr6l7" (OuterVolumeSpecName: "kube-api-access-tr6l7") pod "f529970a-5e33-481e-9b68-4b0519947798" (UID: "f529970a-5e33-481e-9b68-4b0519947798"). InnerVolumeSpecName "kube-api-access-tr6l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.151094 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f529970a-5e33-481e-9b68-4b0519947798" (UID: "f529970a-5e33-481e-9b68-4b0519947798"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.154812 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-config-data" (OuterVolumeSpecName: "config-data") pod "f529970a-5e33-481e-9b68-4b0519947798" (UID: "f529970a-5e33-481e-9b68-4b0519947798"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.188487 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b0778b82-d141-47f3-b766-a3df63f09d12","Type":"ContainerStarted","Data":"0c209afc1d6a0ec35149bb5f607fe3393832b696506b0db4291a508e9b799bb9"} Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.188820 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.190128 4810 generic.go:334] "Generic (PLEG): container finished" podID="f529970a-5e33-481e-9b68-4b0519947798" containerID="4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca" exitCode=0 Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.190183 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.190191 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f529970a-5e33-481e-9b68-4b0519947798","Type":"ContainerDied","Data":"4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca"} Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.190277 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f529970a-5e33-481e-9b68-4b0519947798","Type":"ContainerDied","Data":"23d1dc977cadc63d31e4457bdb9ace8cdf2d8037911fde87d8bbd0f14d3f26a8"} Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.190313 4810 scope.go:117] "RemoveContainer" containerID="4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.209892 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.745248916 podStartE2EDuration="3.209871224s" podCreationTimestamp="2025-09-30 08:24:35 +0000 UTC" firstStartedPulling="2025-09-30 08:24:36.083400525 +0000 UTC m=+1299.535599792" lastFinishedPulling="2025-09-30 08:24:37.548022823 +0000 UTC m=+1301.000222100" observedRunningTime="2025-09-30 08:24:38.208489373 +0000 UTC m=+1301.660688650" watchObservedRunningTime="2025-09-30 08:24:38.209871224 +0000 UTC m=+1301.662070491" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.223352 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.223384 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr6l7\" (UniqueName: \"kubernetes.io/projected/f529970a-5e33-481e-9b68-4b0519947798-kube-api-access-tr6l7\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.223395 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f529970a-5e33-481e-9b68-4b0519947798-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.223403 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f529970a-5e33-481e-9b68-4b0519947798-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.242363 4810 scope.go:117] "RemoveContainer" containerID="a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.244618 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.263780 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.275476 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 08:24:38 crc kubenswrapper[4810]: E0930 08:24:38.275969 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f529970a-5e33-481e-9b68-4b0519947798" containerName="nova-api-api" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.275987 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f529970a-5e33-481e-9b68-4b0519947798" containerName="nova-api-api" Sep 30 08:24:38 crc kubenswrapper[4810]: E0930 08:24:38.276016 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f529970a-5e33-481e-9b68-4b0519947798" containerName="nova-api-log" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.276024 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f529970a-5e33-481e-9b68-4b0519947798" containerName="nova-api-log" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.276243 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f529970a-5e33-481e-9b68-4b0519947798" containerName="nova-api-api" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.276276 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f529970a-5e33-481e-9b68-4b0519947798" containerName="nova-api-log" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.278403 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.282286 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.289706 4810 scope.go:117] "RemoveContainer" containerID="4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.289953 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 08:24:38 crc kubenswrapper[4810]: E0930 08:24:38.298987 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca\": container with ID starting with 4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca not found: ID does not exist" containerID="4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.299051 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca"} err="failed to get container status \"4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca\": rpc error: code = NotFound desc = could not find container \"4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca\": container with ID starting with 4e70ddf8e4a63ba31cf71b87a1c73533f33bd74a616c0d7a650c8805c7e00eca not found: ID does not exist" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.299080 4810 scope.go:117] "RemoveContainer" containerID="a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc" Sep 30 08:24:38 crc kubenswrapper[4810]: E0930 08:24:38.310407 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc\": container with ID starting with a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc not found: ID does not exist" containerID="a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.310446 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc"} err="failed to get container status \"a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc\": rpc error: code = NotFound desc = could not find container \"a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc\": container with ID starting with a4154ba485e74a476ab67556c3cd3448dc1a88a336db9a841fa2ddedf70293bc not found: ID does not exist" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.432901 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks7zl\" (UniqueName: \"kubernetes.io/projected/6ed10177-366b-4be3-b26a-f0c88841eb09-kube-api-access-ks7zl\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.432963 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-config-data\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.433169 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed10177-366b-4be3-b26a-f0c88841eb09-logs\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.433288 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.534343 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.534450 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks7zl\" (UniqueName: \"kubernetes.io/projected/6ed10177-366b-4be3-b26a-f0c88841eb09-kube-api-access-ks7zl\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.534489 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-config-data\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.534542 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed10177-366b-4be3-b26a-f0c88841eb09-logs\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.534949 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed10177-366b-4be3-b26a-f0c88841eb09-logs\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.550237 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-config-data\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.557220 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks7zl\" (UniqueName: \"kubernetes.io/projected/6ed10177-366b-4be3-b26a-f0c88841eb09-kube-api-access-ks7zl\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.560936 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.629090 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.720543 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.845089 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-config-data\") pod \"528064bb-fbc6-42fb-b46d-e932ad6971b1\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.845191 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-scripts\") pod \"528064bb-fbc6-42fb-b46d-e932ad6971b1\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.845230 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-combined-ca-bundle\") pod \"528064bb-fbc6-42fb-b46d-e932ad6971b1\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.845254 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmqfr\" (UniqueName: \"kubernetes.io/projected/528064bb-fbc6-42fb-b46d-e932ad6971b1-kube-api-access-pmqfr\") pod \"528064bb-fbc6-42fb-b46d-e932ad6971b1\" (UID: \"528064bb-fbc6-42fb-b46d-e932ad6971b1\") " Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.850682 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-scripts" (OuterVolumeSpecName: "scripts") pod "528064bb-fbc6-42fb-b46d-e932ad6971b1" (UID: "528064bb-fbc6-42fb-b46d-e932ad6971b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.851352 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/528064bb-fbc6-42fb-b46d-e932ad6971b1-kube-api-access-pmqfr" (OuterVolumeSpecName: "kube-api-access-pmqfr") pod "528064bb-fbc6-42fb-b46d-e932ad6971b1" (UID: "528064bb-fbc6-42fb-b46d-e932ad6971b1"). InnerVolumeSpecName "kube-api-access-pmqfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.895472 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "528064bb-fbc6-42fb-b46d-e932ad6971b1" (UID: "528064bb-fbc6-42fb-b46d-e932ad6971b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.902249 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-config-data" (OuterVolumeSpecName: "config-data") pod "528064bb-fbc6-42fb-b46d-e932ad6971b1" (UID: "528064bb-fbc6-42fb-b46d-e932ad6971b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.949065 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.949111 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.949125 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528064bb-fbc6-42fb-b46d-e932ad6971b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:38 crc kubenswrapper[4810]: I0930 08:24:38.949138 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmqfr\" (UniqueName: \"kubernetes.io/projected/528064bb-fbc6-42fb-b46d-e932ad6971b1-kube-api-access-pmqfr\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.074052 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:24:39 crc kubenswrapper[4810]: W0930 08:24:39.075425 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ed10177_366b_4be3_b26a_f0c88841eb09.slice/crio-d3083fc8db689be10b9d00a47018e0816a0143f0c99dcda294dfd71065d58997 WatchSource:0}: Error finding container d3083fc8db689be10b9d00a47018e0816a0143f0c99dcda294dfd71065d58997: Status 404 returned error can't find the container with id d3083fc8db689be10b9d00a47018e0816a0143f0c99dcda294dfd71065d58997 Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.205638 4810 generic.go:334] "Generic (PLEG): container finished" podID="fc34893b-bbff-4ea5-8c03-57a392f4a788" containerID="393f71ec03f46749da16b49dc6f6ef9b4d126c88d6c8c1fdb8db6a0af09313ca" exitCode=0 Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.205745 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fc34893b-bbff-4ea5-8c03-57a392f4a788","Type":"ContainerDied","Data":"393f71ec03f46749da16b49dc6f6ef9b4d126c88d6c8c1fdb8db6a0af09313ca"} Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.209196 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7l4sp" event={"ID":"528064bb-fbc6-42fb-b46d-e932ad6971b1","Type":"ContainerDied","Data":"9c1128315f58b24fb99fb5eb3f8e34975c2c6a9bd924219886c5cc218cfdf910"} Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.209215 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c1128315f58b24fb99fb5eb3f8e34975c2c6a9bd924219886c5cc218cfdf910" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.209237 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7l4sp" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.211666 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ed10177-366b-4be3-b26a-f0c88841eb09","Type":"ContainerStarted","Data":"d3083fc8db689be10b9d00a47018e0816a0143f0c99dcda294dfd71065d58997"} Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.262414 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 08:24:39 crc kubenswrapper[4810]: E0930 08:24:39.264595 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528064bb-fbc6-42fb-b46d-e932ad6971b1" containerName="nova-cell1-conductor-db-sync" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.267864 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="528064bb-fbc6-42fb-b46d-e932ad6971b1" containerName="nova-cell1-conductor-db-sync" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.268803 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="528064bb-fbc6-42fb-b46d-e932ad6971b1" containerName="nova-cell1-conductor-db-sync" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.269682 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.275177 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.282723 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.297663 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.331790 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f529970a-5e33-481e-9b68-4b0519947798" path="/var/lib/kubelet/pods/f529970a-5e33-481e-9b68-4b0519947798/volumes" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.371677 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-combined-ca-bundle\") pod \"fc34893b-bbff-4ea5-8c03-57a392f4a788\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.371981 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcrbq\" (UniqueName: \"kubernetes.io/projected/fc34893b-bbff-4ea5-8c03-57a392f4a788-kube-api-access-rcrbq\") pod \"fc34893b-bbff-4ea5-8c03-57a392f4a788\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.372036 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-config-data\") pod \"fc34893b-bbff-4ea5-8c03-57a392f4a788\" (UID: \"fc34893b-bbff-4ea5-8c03-57a392f4a788\") " Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.372368 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrp6n\" (UniqueName: \"kubernetes.io/projected/4342c2f0-9d27-4211-a302-a1822255f011-kube-api-access-hrp6n\") pod \"nova-cell1-conductor-0\" (UID: \"4342c2f0-9d27-4211-a302-a1822255f011\") " pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.372447 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4342c2f0-9d27-4211-a302-a1822255f011-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4342c2f0-9d27-4211-a302-a1822255f011\") " pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.372474 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4342c2f0-9d27-4211-a302-a1822255f011-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4342c2f0-9d27-4211-a302-a1822255f011\") " pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.378414 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc34893b-bbff-4ea5-8c03-57a392f4a788-kube-api-access-rcrbq" (OuterVolumeSpecName: "kube-api-access-rcrbq") pod "fc34893b-bbff-4ea5-8c03-57a392f4a788" (UID: "fc34893b-bbff-4ea5-8c03-57a392f4a788"). InnerVolumeSpecName "kube-api-access-rcrbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.397891 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-config-data" (OuterVolumeSpecName: "config-data") pod "fc34893b-bbff-4ea5-8c03-57a392f4a788" (UID: "fc34893b-bbff-4ea5-8c03-57a392f4a788"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.419166 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc34893b-bbff-4ea5-8c03-57a392f4a788" (UID: "fc34893b-bbff-4ea5-8c03-57a392f4a788"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.476126 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4342c2f0-9d27-4211-a302-a1822255f011-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4342c2f0-9d27-4211-a302-a1822255f011\") " pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.477118 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4342c2f0-9d27-4211-a302-a1822255f011-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4342c2f0-9d27-4211-a302-a1822255f011\") " pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.478309 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrp6n\" (UniqueName: \"kubernetes.io/projected/4342c2f0-9d27-4211-a302-a1822255f011-kube-api-access-hrp6n\") pod \"nova-cell1-conductor-0\" (UID: \"4342c2f0-9d27-4211-a302-a1822255f011\") " pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.483167 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4342c2f0-9d27-4211-a302-a1822255f011-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4342c2f0-9d27-4211-a302-a1822255f011\") " pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.483379 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4342c2f0-9d27-4211-a302-a1822255f011-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4342c2f0-9d27-4211-a302-a1822255f011\") " pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.483536 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcrbq\" (UniqueName: \"kubernetes.io/projected/fc34893b-bbff-4ea5-8c03-57a392f4a788-kube-api-access-rcrbq\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.483566 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.483578 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc34893b-bbff-4ea5-8c03-57a392f4a788-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.495420 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrp6n\" (UniqueName: \"kubernetes.io/projected/4342c2f0-9d27-4211-a302-a1822255f011-kube-api-access-hrp6n\") pod \"nova-cell1-conductor-0\" (UID: \"4342c2f0-9d27-4211-a302-a1822255f011\") " pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:39 crc kubenswrapper[4810]: I0930 08:24:39.617739 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:40 crc kubenswrapper[4810]: W0930 08:24:40.100831 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4342c2f0_9d27_4211_a302_a1822255f011.slice/crio-24aaaf0a89762bb3ca806c77cf2326bf1a8a4a15a9cf7255d45c52464c0e2933 WatchSource:0}: Error finding container 24aaaf0a89762bb3ca806c77cf2326bf1a8a4a15a9cf7255d45c52464c0e2933: Status 404 returned error can't find the container with id 24aaaf0a89762bb3ca806c77cf2326bf1a8a4a15a9cf7255d45c52464c0e2933 Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.101325 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.226757 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ed10177-366b-4be3-b26a-f0c88841eb09","Type":"ContainerStarted","Data":"c8a6f55db2bed56e970121f4fd7ef18f7da3fde6130b6be3f31b65d0967d886b"} Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.227164 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ed10177-366b-4be3-b26a-f0c88841eb09","Type":"ContainerStarted","Data":"56263de56fc274b57866a8ff801e7a382a9472f40897e29754ddd36ab39b25ba"} Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.229242 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.229247 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fc34893b-bbff-4ea5-8c03-57a392f4a788","Type":"ContainerDied","Data":"4398d52bafeb96d005e05ece446279a87632ea329cac84847dd6f503fb902ce5"} Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.229479 4810 scope.go:117] "RemoveContainer" containerID="393f71ec03f46749da16b49dc6f6ef9b4d126c88d6c8c1fdb8db6a0af09313ca" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.231833 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4342c2f0-9d27-4211-a302-a1822255f011","Type":"ContainerStarted","Data":"24aaaf0a89762bb3ca806c77cf2326bf1a8a4a15a9cf7255d45c52464c0e2933"} Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.263336 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.263316141 podStartE2EDuration="2.263316141s" podCreationTimestamp="2025-09-30 08:24:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:24:40.258429119 +0000 UTC m=+1303.710628396" watchObservedRunningTime="2025-09-30 08:24:40.263316141 +0000 UTC m=+1303.715515408" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.280505 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.290769 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.298786 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:24:40 crc kubenswrapper[4810]: E0930 08:24:40.299205 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc34893b-bbff-4ea5-8c03-57a392f4a788" containerName="nova-scheduler-scheduler" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.299223 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc34893b-bbff-4ea5-8c03-57a392f4a788" containerName="nova-scheduler-scheduler" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.299426 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc34893b-bbff-4ea5-8c03-57a392f4a788" containerName="nova-scheduler-scheduler" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.300092 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.306528 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.310255 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.403122 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.403179 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-config-data\") pod \"nova-scheduler-0\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.404573 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4j9r\" (UniqueName: \"kubernetes.io/projected/60aa73ab-48bd-4955-a69e-ef1993fa420e-kube-api-access-h4j9r\") pod \"nova-scheduler-0\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.507186 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4j9r\" (UniqueName: \"kubernetes.io/projected/60aa73ab-48bd-4955-a69e-ef1993fa420e-kube-api-access-h4j9r\") pod \"nova-scheduler-0\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.507330 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.507402 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-config-data\") pod \"nova-scheduler-0\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.513693 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-config-data\") pod \"nova-scheduler-0\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.514548 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.541064 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4j9r\" (UniqueName: \"kubernetes.io/projected/60aa73ab-48bd-4955-a69e-ef1993fa420e-kube-api-access-h4j9r\") pod \"nova-scheduler-0\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " pod="openstack/nova-scheduler-0" Sep 30 08:24:40 crc kubenswrapper[4810]: I0930 08:24:40.617700 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 08:24:41 crc kubenswrapper[4810]: I0930 08:24:41.137339 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:24:41 crc kubenswrapper[4810]: W0930 08:24:41.143750 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60aa73ab_48bd_4955_a69e_ef1993fa420e.slice/crio-a8c8904e8eede7adc0e24234724d5769a1896ea8bbbd7903e2f64ee32ca8f1a8 WatchSource:0}: Error finding container a8c8904e8eede7adc0e24234724d5769a1896ea8bbbd7903e2f64ee32ca8f1a8: Status 404 returned error can't find the container with id a8c8904e8eede7adc0e24234724d5769a1896ea8bbbd7903e2f64ee32ca8f1a8 Sep 30 08:24:41 crc kubenswrapper[4810]: I0930 08:24:41.247679 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4342c2f0-9d27-4211-a302-a1822255f011","Type":"ContainerStarted","Data":"813eeda4e604e9e45bc3355490945985e0b377bf6c0232c42ae981963fe67650"} Sep 30 08:24:41 crc kubenswrapper[4810]: I0930 08:24:41.248180 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:41 crc kubenswrapper[4810]: I0930 08:24:41.250519 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60aa73ab-48bd-4955-a69e-ef1993fa420e","Type":"ContainerStarted","Data":"a8c8904e8eede7adc0e24234724d5769a1896ea8bbbd7903e2f64ee32ca8f1a8"} Sep 30 08:24:41 crc kubenswrapper[4810]: I0930 08:24:41.268755 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.268737365 podStartE2EDuration="2.268737365s" podCreationTimestamp="2025-09-30 08:24:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:24:41.265819421 +0000 UTC m=+1304.718018688" watchObservedRunningTime="2025-09-30 08:24:41.268737365 +0000 UTC m=+1304.720936632" Sep 30 08:24:41 crc kubenswrapper[4810]: I0930 08:24:41.318992 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc34893b-bbff-4ea5-8c03-57a392f4a788" path="/var/lib/kubelet/pods/fc34893b-bbff-4ea5-8c03-57a392f4a788/volumes" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.265307 4810 generic.go:334] "Generic (PLEG): container finished" podID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerID="4b0eb2e8c789b768a3b2b5204b4e1f11b0ddae88b366291c94522d464110a616" exitCode=0 Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.265480 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3","Type":"ContainerDied","Data":"4b0eb2e8c789b768a3b2b5204b4e1f11b0ddae88b366291c94522d464110a616"} Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.267553 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60aa73ab-48bd-4955-a69e-ef1993fa420e","Type":"ContainerStarted","Data":"53c240cb16bb6f14f8acc07cc944df9a8fe30ec0f31adbd2c28781ac6153c810"} Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.301583 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.301551045 podStartE2EDuration="2.301551045s" podCreationTimestamp="2025-09-30 08:24:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:24:42.282882933 +0000 UTC m=+1305.735082230" watchObservedRunningTime="2025-09-30 08:24:42.301551045 +0000 UTC m=+1305.753750312" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.691114 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.851087 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-combined-ca-bundle\") pod \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.851175 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-config-data\") pod \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.851295 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-log-httpd\") pod \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.851324 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-sg-core-conf-yaml\") pod \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.851410 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-run-httpd\") pod \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.851553 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9n99\" (UniqueName: \"kubernetes.io/projected/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-kube-api-access-t9n99\") pod \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.851614 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-scripts\") pod \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\" (UID: \"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3\") " Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.852224 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" (UID: "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.853194 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" (UID: "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.860478 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-kube-api-access-t9n99" (OuterVolumeSpecName: "kube-api-access-t9n99") pod "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" (UID: "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3"). InnerVolumeSpecName "kube-api-access-t9n99". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.860626 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-scripts" (OuterVolumeSpecName: "scripts") pod "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" (UID: "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.886198 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" (UID: "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.945783 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" (UID: "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.955485 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9n99\" (UniqueName: \"kubernetes.io/projected/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-kube-api-access-t9n99\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.955546 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.955560 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.955571 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.955581 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.955591 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:42 crc kubenswrapper[4810]: I0930 08:24:42.979034 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-config-data" (OuterVolumeSpecName: "config-data") pod "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" (UID: "6372d9e0-5994-4f45-8e6c-d126c9bfbdc3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.057506 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.281464 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6372d9e0-5994-4f45-8e6c-d126c9bfbdc3","Type":"ContainerDied","Data":"4c4f299bb218fcbdc5b7b60e0952ecf3a4a0fdac7adee4c90e2587f7618e51f8"} Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.281757 4810 scope.go:117] "RemoveContainer" containerID="4d05bb820cf7cd66583e21de9f009f14c384a982e01de30df61b96d413e2cb89" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.281521 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.300868 4810 scope.go:117] "RemoveContainer" containerID="0cdcfe02e4b69f677ff022b9d8c8f704147d4b38d2336beb526e85f05ba447ac" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.338609 4810 scope.go:117] "RemoveContainer" containerID="4b0eb2e8c789b768a3b2b5204b4e1f11b0ddae88b366291c94522d464110a616" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.354916 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.367794 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.371954 4810 scope.go:117] "RemoveContainer" containerID="2ea9647b347e72f0da13d4bcc352258e5241434ca899960e013eb65abfb2de42" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.387045 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:24:43 crc kubenswrapper[4810]: E0930 08:24:43.387794 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="proxy-httpd" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.387827 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="proxy-httpd" Sep 30 08:24:43 crc kubenswrapper[4810]: E0930 08:24:43.387871 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="ceilometer-notification-agent" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.387881 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="ceilometer-notification-agent" Sep 30 08:24:43 crc kubenswrapper[4810]: E0930 08:24:43.387897 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="ceilometer-central-agent" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.387907 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="ceilometer-central-agent" Sep 30 08:24:43 crc kubenswrapper[4810]: E0930 08:24:43.387932 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="sg-core" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.387940 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="sg-core" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.388167 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="proxy-httpd" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.388200 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="ceilometer-central-agent" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.388213 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="ceilometer-notification-agent" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.388235 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" containerName="sg-core" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.390576 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.395057 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.415677 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.415899 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.416052 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.568636 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-run-httpd\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.568688 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-scripts\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.568723 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.569154 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.569253 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-config-data\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.569406 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hrbw\" (UniqueName: \"kubernetes.io/projected/825a301d-094e-4c79-8b68-2f038635df38-kube-api-access-7hrbw\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.569735 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-log-httpd\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.569857 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.671094 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.671140 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-config-data\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.671172 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hrbw\" (UniqueName: \"kubernetes.io/projected/825a301d-094e-4c79-8b68-2f038635df38-kube-api-access-7hrbw\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.671226 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-log-httpd\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.671248 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.671320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-run-httpd\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.671351 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-scripts\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.671387 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.672664 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-run-httpd\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.672839 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-log-httpd\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.678116 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.679439 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-config-data\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.679955 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-scripts\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.697801 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.700846 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.701671 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hrbw\" (UniqueName: \"kubernetes.io/projected/825a301d-094e-4c79-8b68-2f038635df38-kube-api-access-7hrbw\") pod \"ceilometer-0\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " pod="openstack/ceilometer-0" Sep 30 08:24:43 crc kubenswrapper[4810]: I0930 08:24:43.740769 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:24:44 crc kubenswrapper[4810]: I0930 08:24:44.224346 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:24:44 crc kubenswrapper[4810]: W0930 08:24:44.226573 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod825a301d_094e_4c79_8b68_2f038635df38.slice/crio-c65d54f8ac0ec44dff3b04865c6a9a30210adabb4ca3803d5f7aae5bff4684c4 WatchSource:0}: Error finding container c65d54f8ac0ec44dff3b04865c6a9a30210adabb4ca3803d5f7aae5bff4684c4: Status 404 returned error can't find the container with id c65d54f8ac0ec44dff3b04865c6a9a30210adabb4ca3803d5f7aae5bff4684c4 Sep 30 08:24:44 crc kubenswrapper[4810]: I0930 08:24:44.293510 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"825a301d-094e-4c79-8b68-2f038635df38","Type":"ContainerStarted","Data":"c65d54f8ac0ec44dff3b04865c6a9a30210adabb4ca3803d5f7aae5bff4684c4"} Sep 30 08:24:45 crc kubenswrapper[4810]: I0930 08:24:45.326201 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6372d9e0-5994-4f45-8e6c-d126c9bfbdc3" path="/var/lib/kubelet/pods/6372d9e0-5994-4f45-8e6c-d126c9bfbdc3/volumes" Sep 30 08:24:45 crc kubenswrapper[4810]: I0930 08:24:45.328238 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"825a301d-094e-4c79-8b68-2f038635df38","Type":"ContainerStarted","Data":"2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82"} Sep 30 08:24:45 crc kubenswrapper[4810]: I0930 08:24:45.328345 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"825a301d-094e-4c79-8b68-2f038635df38","Type":"ContainerStarted","Data":"24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e"} Sep 30 08:24:45 crc kubenswrapper[4810]: I0930 08:24:45.570939 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 30 08:24:45 crc kubenswrapper[4810]: I0930 08:24:45.617984 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 08:24:46 crc kubenswrapper[4810]: I0930 08:24:46.321004 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"825a301d-094e-4c79-8b68-2f038635df38","Type":"ContainerStarted","Data":"74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f"} Sep 30 08:24:48 crc kubenswrapper[4810]: I0930 08:24:48.350579 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"825a301d-094e-4c79-8b68-2f038635df38","Type":"ContainerStarted","Data":"f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8"} Sep 30 08:24:48 crc kubenswrapper[4810]: I0930 08:24:48.351211 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 08:24:48 crc kubenswrapper[4810]: I0930 08:24:48.381020 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.023936569 podStartE2EDuration="5.381000475s" podCreationTimestamp="2025-09-30 08:24:43 +0000 UTC" firstStartedPulling="2025-09-30 08:24:44.229040832 +0000 UTC m=+1307.681240099" lastFinishedPulling="2025-09-30 08:24:47.586104728 +0000 UTC m=+1311.038304005" observedRunningTime="2025-09-30 08:24:48.378951556 +0000 UTC m=+1311.831150823" watchObservedRunningTime="2025-09-30 08:24:48.381000475 +0000 UTC m=+1311.833199742" Sep 30 08:24:48 crc kubenswrapper[4810]: I0930 08:24:48.629849 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 08:24:48 crc kubenswrapper[4810]: I0930 08:24:48.631326 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 08:24:49 crc kubenswrapper[4810]: I0930 08:24:49.653466 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Sep 30 08:24:49 crc kubenswrapper[4810]: I0930 08:24:49.713416 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.216:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 08:24:49 crc kubenswrapper[4810]: I0930 08:24:49.713500 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.216:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 08:24:50 crc kubenswrapper[4810]: I0930 08:24:50.618444 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 08:24:50 crc kubenswrapper[4810]: I0930 08:24:50.654932 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 08:24:51 crc kubenswrapper[4810]: I0930 08:24:51.408382 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.461812 4810 generic.go:334] "Generic (PLEG): container finished" podID="17a8b230-cfc5-415e-a480-d78e8bdff63c" containerID="e0f948ff00373a2e51bfff5be02da66a3df8c5dcf94f95481b9b6b625fa15d65" exitCode=137 Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.462433 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17a8b230-cfc5-415e-a480-d78e8bdff63c","Type":"ContainerDied","Data":"e0f948ff00373a2e51bfff5be02da66a3df8c5dcf94f95481b9b6b625fa15d65"} Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.469937 4810 generic.go:334] "Generic (PLEG): container finished" podID="17299686-04ed-4de5-b64c-a504a5783e83" containerID="a6a3507362cecedd81e8b65e684e774019c2c108eb2c9d6e8f2010b460be5ba0" exitCode=137 Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.469982 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"17299686-04ed-4de5-b64c-a504a5783e83","Type":"ContainerDied","Data":"a6a3507362cecedd81e8b65e684e774019c2c108eb2c9d6e8f2010b460be5ba0"} Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.635326 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.643698 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.795204 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5dq2\" (UniqueName: \"kubernetes.io/projected/17a8b230-cfc5-415e-a480-d78e8bdff63c-kube-api-access-v5dq2\") pod \"17a8b230-cfc5-415e-a480-d78e8bdff63c\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.795285 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzptp\" (UniqueName: \"kubernetes.io/projected/17299686-04ed-4de5-b64c-a504a5783e83-kube-api-access-jzptp\") pod \"17299686-04ed-4de5-b64c-a504a5783e83\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.795343 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-combined-ca-bundle\") pod \"17a8b230-cfc5-415e-a480-d78e8bdff63c\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.795379 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-combined-ca-bundle\") pod \"17299686-04ed-4de5-b64c-a504a5783e83\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.795441 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17a8b230-cfc5-415e-a480-d78e8bdff63c-logs\") pod \"17a8b230-cfc5-415e-a480-d78e8bdff63c\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.795516 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-config-data\") pod \"17a8b230-cfc5-415e-a480-d78e8bdff63c\" (UID: \"17a8b230-cfc5-415e-a480-d78e8bdff63c\") " Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.795537 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-config-data\") pod \"17299686-04ed-4de5-b64c-a504a5783e83\" (UID: \"17299686-04ed-4de5-b64c-a504a5783e83\") " Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.795943 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17a8b230-cfc5-415e-a480-d78e8bdff63c-logs" (OuterVolumeSpecName: "logs") pod "17a8b230-cfc5-415e-a480-d78e8bdff63c" (UID: "17a8b230-cfc5-415e-a480-d78e8bdff63c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.801639 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17a8b230-cfc5-415e-a480-d78e8bdff63c-kube-api-access-v5dq2" (OuterVolumeSpecName: "kube-api-access-v5dq2") pod "17a8b230-cfc5-415e-a480-d78e8bdff63c" (UID: "17a8b230-cfc5-415e-a480-d78e8bdff63c"). InnerVolumeSpecName "kube-api-access-v5dq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.802366 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17299686-04ed-4de5-b64c-a504a5783e83-kube-api-access-jzptp" (OuterVolumeSpecName: "kube-api-access-jzptp") pod "17299686-04ed-4de5-b64c-a504a5783e83" (UID: "17299686-04ed-4de5-b64c-a504a5783e83"). InnerVolumeSpecName "kube-api-access-jzptp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.829708 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17299686-04ed-4de5-b64c-a504a5783e83" (UID: "17299686-04ed-4de5-b64c-a504a5783e83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.839427 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17a8b230-cfc5-415e-a480-d78e8bdff63c" (UID: "17a8b230-cfc5-415e-a480-d78e8bdff63c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.839957 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-config-data" (OuterVolumeSpecName: "config-data") pod "17a8b230-cfc5-415e-a480-d78e8bdff63c" (UID: "17a8b230-cfc5-415e-a480-d78e8bdff63c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.849749 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-config-data" (OuterVolumeSpecName: "config-data") pod "17299686-04ed-4de5-b64c-a504a5783e83" (UID: "17299686-04ed-4de5-b64c-a504a5783e83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.897962 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5dq2\" (UniqueName: \"kubernetes.io/projected/17a8b230-cfc5-415e-a480-d78e8bdff63c-kube-api-access-v5dq2\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.898002 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzptp\" (UniqueName: \"kubernetes.io/projected/17299686-04ed-4de5-b64c-a504a5783e83-kube-api-access-jzptp\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.898018 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.898030 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.898041 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17a8b230-cfc5-415e-a480-d78e8bdff63c-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.898056 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17a8b230-cfc5-415e-a480-d78e8bdff63c-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:57 crc kubenswrapper[4810]: I0930 08:24:57.898066 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17299686-04ed-4de5-b64c-a504a5783e83-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.493387 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"17299686-04ed-4de5-b64c-a504a5783e83","Type":"ContainerDied","Data":"02281b3a1b6b65c763effe07b64d4436559d36d8b28e9f4c058a197d992d38fd"} Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.493451 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.493470 4810 scope.go:117] "RemoveContainer" containerID="a6a3507362cecedd81e8b65e684e774019c2c108eb2c9d6e8f2010b460be5ba0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.496927 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17a8b230-cfc5-415e-a480-d78e8bdff63c","Type":"ContainerDied","Data":"ecadbf42074cc5637ec261c40f35c3be17a7ed88bf59219018934722eab39459"} Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.496988 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.516963 4810 scope.go:117] "RemoveContainer" containerID="e0f948ff00373a2e51bfff5be02da66a3df8c5dcf94f95481b9b6b625fa15d65" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.552654 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.565943 4810 scope.go:117] "RemoveContainer" containerID="c3662bc080135b038a8b4747a99cc4bf4db82aa7d38b2bcb37cf6066588622b4" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.585249 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.593982 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.608846 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.614149 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 08:24:58 crc kubenswrapper[4810]: E0930 08:24:58.619080 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17299686-04ed-4de5-b64c-a504a5783e83" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.619117 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="17299686-04ed-4de5-b64c-a504a5783e83" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 08:24:58 crc kubenswrapper[4810]: E0930 08:24:58.619355 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17a8b230-cfc5-415e-a480-d78e8bdff63c" containerName="nova-metadata-metadata" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.619367 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="17a8b230-cfc5-415e-a480-d78e8bdff63c" containerName="nova-metadata-metadata" Sep 30 08:24:58 crc kubenswrapper[4810]: E0930 08:24:58.619395 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17a8b230-cfc5-415e-a480-d78e8bdff63c" containerName="nova-metadata-log" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.619404 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="17a8b230-cfc5-415e-a480-d78e8bdff63c" containerName="nova-metadata-log" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.621945 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="17a8b230-cfc5-415e-a480-d78e8bdff63c" containerName="nova-metadata-log" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.621979 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="17a8b230-cfc5-415e-a480-d78e8bdff63c" containerName="nova-metadata-metadata" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.621992 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="17299686-04ed-4de5-b64c-a504a5783e83" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.623310 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.627554 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.628558 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.629508 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.639227 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.642921 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.643621 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.643925 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.643921 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.649683 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.649834 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.649855 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.653725 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.663222 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.715526 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.715586 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ll4j\" (UniqueName: \"kubernetes.io/projected/fc321aeb-240d-414e-91de-ba95671d6556-kube-api-access-6ll4j\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.715608 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.715665 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.715977 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.819067 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-config-data\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.819181 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.819309 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.819592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ll4j\" (UniqueName: \"kubernetes.io/projected/fc321aeb-240d-414e-91de-ba95671d6556-kube-api-access-6ll4j\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.819679 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.819770 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.819935 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be553dcb-b399-458b-8e33-482a0cf48ba7-logs\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.820451 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.820612 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88n29\" (UniqueName: \"kubernetes.io/projected/be553dcb-b399-458b-8e33-482a0cf48ba7-kube-api-access-88n29\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.820706 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.826660 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.827105 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.828102 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.831045 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc321aeb-240d-414e-91de-ba95671d6556-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.853154 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ll4j\" (UniqueName: \"kubernetes.io/projected/fc321aeb-240d-414e-91de-ba95671d6556-kube-api-access-6ll4j\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc321aeb-240d-414e-91de-ba95671d6556\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.922363 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be553dcb-b399-458b-8e33-482a0cf48ba7-logs\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.922492 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.922525 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88n29\" (UniqueName: \"kubernetes.io/projected/be553dcb-b399-458b-8e33-482a0cf48ba7-kube-api-access-88n29\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.922579 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-config-data\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.922608 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.923036 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be553dcb-b399-458b-8e33-482a0cf48ba7-logs\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.926708 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.929375 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-config-data\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.929834 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.939054 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88n29\" (UniqueName: \"kubernetes.io/projected/be553dcb-b399-458b-8e33-482a0cf48ba7-kube-api-access-88n29\") pod \"nova-metadata-0\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " pod="openstack/nova-metadata-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.952278 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:24:58 crc kubenswrapper[4810]: I0930 08:24:58.972223 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.320513 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17299686-04ed-4de5-b64c-a504a5783e83" path="/var/lib/kubelet/pods/17299686-04ed-4de5-b64c-a504a5783e83/volumes" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.321096 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17a8b230-cfc5-415e-a480-d78e8bdff63c" path="/var/lib/kubelet/pods/17a8b230-cfc5-415e-a480-d78e8bdff63c/volumes" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.504354 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.513323 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.523111 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.590506 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.734121 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb9f56f5-cg226"] Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.736028 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.757335 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb9f56f5-cg226"] Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.878061 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-svc\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.878132 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.878232 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.878288 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drnbg\" (UniqueName: \"kubernetes.io/projected/10c33784-e317-418b-bdf5-ea60acd29d98-kube-api-access-drnbg\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.878330 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.878391 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-config\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.980156 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.980532 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-config\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.980570 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-svc\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.980607 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.980656 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.980718 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drnbg\" (UniqueName: \"kubernetes.io/projected/10c33784-e317-418b-bdf5-ea60acd29d98-kube-api-access-drnbg\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.981758 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.983219 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.983427 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-svc\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.983523 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.983743 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-config\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:24:59 crc kubenswrapper[4810]: I0930 08:24:59.997544 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drnbg\" (UniqueName: \"kubernetes.io/projected/10c33784-e317-418b-bdf5-ea60acd29d98-kube-api-access-drnbg\") pod \"dnsmasq-dns-6bb9f56f5-cg226\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:25:00 crc kubenswrapper[4810]: I0930 08:25:00.128772 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:25:00 crc kubenswrapper[4810]: I0930 08:25:00.521673 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fc321aeb-240d-414e-91de-ba95671d6556","Type":"ContainerStarted","Data":"58bcf8122fddc0be6820c3b8eb70101cca3a1fef7c983183b3d0da0e60c42b3d"} Sep 30 08:25:00 crc kubenswrapper[4810]: I0930 08:25:00.521922 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fc321aeb-240d-414e-91de-ba95671d6556","Type":"ContainerStarted","Data":"098d4bf408bb9d495179cc6a9fa442454ebb1d2fe2864118bd644fa0dc532e61"} Sep 30 08:25:00 crc kubenswrapper[4810]: I0930 08:25:00.526249 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be553dcb-b399-458b-8e33-482a0cf48ba7","Type":"ContainerStarted","Data":"b2e64049309fffd0cc128b6123caa07c1ccb1191f677273a1715030c3141fdbd"} Sep 30 08:25:00 crc kubenswrapper[4810]: I0930 08:25:00.526306 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be553dcb-b399-458b-8e33-482a0cf48ba7","Type":"ContainerStarted","Data":"e1004c44a710c6b1586c7ec58d8cb79dfbbe1bc7012bef84028057c07d6549a6"} Sep 30 08:25:00 crc kubenswrapper[4810]: I0930 08:25:00.526324 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be553dcb-b399-458b-8e33-482a0cf48ba7","Type":"ContainerStarted","Data":"d8ed2db606176813feca796e503e2711f07aed7aaf991d923f3ecbc4028efd28"} Sep 30 08:25:00 crc kubenswrapper[4810]: I0930 08:25:00.538957 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.538938076 podStartE2EDuration="2.538938076s" podCreationTimestamp="2025-09-30 08:24:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:25:00.536119254 +0000 UTC m=+1323.988318531" watchObservedRunningTime="2025-09-30 08:25:00.538938076 +0000 UTC m=+1323.991137343" Sep 30 08:25:00 crc kubenswrapper[4810]: I0930 08:25:00.606351 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.606327134 podStartE2EDuration="2.606327134s" podCreationTimestamp="2025-09-30 08:24:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:25:00.563913371 +0000 UTC m=+1324.016112638" watchObservedRunningTime="2025-09-30 08:25:00.606327134 +0000 UTC m=+1324.058526401" Sep 30 08:25:00 crc kubenswrapper[4810]: I0930 08:25:00.613470 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb9f56f5-cg226"] Sep 30 08:25:01 crc kubenswrapper[4810]: I0930 08:25:01.542582 4810 generic.go:334] "Generic (PLEG): container finished" podID="10c33784-e317-418b-bdf5-ea60acd29d98" containerID="0d3d5c82441542254298d6151a1d8291d9d91d20d33a937c53f8c8f60ad88742" exitCode=0 Sep 30 08:25:01 crc kubenswrapper[4810]: I0930 08:25:01.544248 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" event={"ID":"10c33784-e317-418b-bdf5-ea60acd29d98","Type":"ContainerDied","Data":"0d3d5c82441542254298d6151a1d8291d9d91d20d33a937c53f8c8f60ad88742"} Sep 30 08:25:01 crc kubenswrapper[4810]: I0930 08:25:01.544295 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" event={"ID":"10c33784-e317-418b-bdf5-ea60acd29d98","Type":"ContainerStarted","Data":"4417d457ba459e14102f801da08e5dc17d73201dd10f265a6ffb902212d4edbd"} Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.047459 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.047851 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="proxy-httpd" containerID="cri-o://f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8" gracePeriod=30 Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.047913 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="sg-core" containerID="cri-o://74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f" gracePeriod=30 Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.048006 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="ceilometer-notification-agent" containerID="cri-o://2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82" gracePeriod=30 Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.048076 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="ceilometer-central-agent" containerID="cri-o://24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e" gracePeriod=30 Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.062508 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.219:3000/\": read tcp 10.217.0.2:57486->10.217.0.219:3000: read: connection reset by peer" Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.328763 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.554986 4810 generic.go:334] "Generic (PLEG): container finished" podID="825a301d-094e-4c79-8b68-2f038635df38" containerID="f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8" exitCode=0 Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.555023 4810 generic.go:334] "Generic (PLEG): container finished" podID="825a301d-094e-4c79-8b68-2f038635df38" containerID="74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f" exitCode=2 Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.555041 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"825a301d-094e-4c79-8b68-2f038635df38","Type":"ContainerDied","Data":"f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8"} Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.555095 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"825a301d-094e-4c79-8b68-2f038635df38","Type":"ContainerDied","Data":"74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f"} Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.557252 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerName="nova-api-log" containerID="cri-o://56263de56fc274b57866a8ff801e7a382a9472f40897e29754ddd36ab39b25ba" gracePeriod=30 Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.557331 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" event={"ID":"10c33784-e317-418b-bdf5-ea60acd29d98","Type":"ContainerStarted","Data":"24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee"} Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.557409 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerName="nova-api-api" containerID="cri-o://c8a6f55db2bed56e970121f4fd7ef18f7da3fde6130b6be3f31b65d0967d886b" gracePeriod=30 Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.557767 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:25:02 crc kubenswrapper[4810]: I0930 08:25:02.590091 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" podStartSLOduration=3.590069636 podStartE2EDuration="3.590069636s" podCreationTimestamp="2025-09-30 08:24:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:25:02.581180267 +0000 UTC m=+1326.033379534" watchObservedRunningTime="2025-09-30 08:25:02.590069636 +0000 UTC m=+1326.042268903" Sep 30 08:25:03 crc kubenswrapper[4810]: I0930 08:25:03.573995 4810 generic.go:334] "Generic (PLEG): container finished" podID="825a301d-094e-4c79-8b68-2f038635df38" containerID="24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e" exitCode=0 Sep 30 08:25:03 crc kubenswrapper[4810]: I0930 08:25:03.574330 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"825a301d-094e-4c79-8b68-2f038635df38","Type":"ContainerDied","Data":"24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e"} Sep 30 08:25:03 crc kubenswrapper[4810]: I0930 08:25:03.576335 4810 generic.go:334] "Generic (PLEG): container finished" podID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerID="c8a6f55db2bed56e970121f4fd7ef18f7da3fde6130b6be3f31b65d0967d886b" exitCode=0 Sep 30 08:25:03 crc kubenswrapper[4810]: I0930 08:25:03.576359 4810 generic.go:334] "Generic (PLEG): container finished" podID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerID="56263de56fc274b57866a8ff801e7a382a9472f40897e29754ddd36ab39b25ba" exitCode=143 Sep 30 08:25:03 crc kubenswrapper[4810]: I0930 08:25:03.577323 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ed10177-366b-4be3-b26a-f0c88841eb09","Type":"ContainerDied","Data":"c8a6f55db2bed56e970121f4fd7ef18f7da3fde6130b6be3f31b65d0967d886b"} Sep 30 08:25:03 crc kubenswrapper[4810]: I0930 08:25:03.577397 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ed10177-366b-4be3-b26a-f0c88841eb09","Type":"ContainerDied","Data":"56263de56fc274b57866a8ff801e7a382a9472f40897e29754ddd36ab39b25ba"} Sep 30 08:25:03 crc kubenswrapper[4810]: I0930 08:25:03.952432 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:25:03 crc kubenswrapper[4810]: I0930 08:25:03.954168 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:25:03 crc kubenswrapper[4810]: I0930 08:25:03.972864 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 08:25:03 crc kubenswrapper[4810]: I0930 08:25:03.973168 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.115574 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-combined-ca-bundle\") pod \"6ed10177-366b-4be3-b26a-f0c88841eb09\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.115802 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed10177-366b-4be3-b26a-f0c88841eb09-logs\") pod \"6ed10177-366b-4be3-b26a-f0c88841eb09\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.115875 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-config-data\") pod \"6ed10177-366b-4be3-b26a-f0c88841eb09\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.115909 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks7zl\" (UniqueName: \"kubernetes.io/projected/6ed10177-366b-4be3-b26a-f0c88841eb09-kube-api-access-ks7zl\") pod \"6ed10177-366b-4be3-b26a-f0c88841eb09\" (UID: \"6ed10177-366b-4be3-b26a-f0c88841eb09\") " Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.119739 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ed10177-366b-4be3-b26a-f0c88841eb09-logs" (OuterVolumeSpecName: "logs") pod "6ed10177-366b-4be3-b26a-f0c88841eb09" (UID: "6ed10177-366b-4be3-b26a-f0c88841eb09"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.132083 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ed10177-366b-4be3-b26a-f0c88841eb09-kube-api-access-ks7zl" (OuterVolumeSpecName: "kube-api-access-ks7zl") pod "6ed10177-366b-4be3-b26a-f0c88841eb09" (UID: "6ed10177-366b-4be3-b26a-f0c88841eb09"). InnerVolumeSpecName "kube-api-access-ks7zl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.156377 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-config-data" (OuterVolumeSpecName: "config-data") pod "6ed10177-366b-4be3-b26a-f0c88841eb09" (UID: "6ed10177-366b-4be3-b26a-f0c88841eb09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.165736 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ed10177-366b-4be3-b26a-f0c88841eb09" (UID: "6ed10177-366b-4be3-b26a-f0c88841eb09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.218438 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.218473 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed10177-366b-4be3-b26a-f0c88841eb09-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.218483 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed10177-366b-4be3-b26a-f0c88841eb09-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.218491 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks7zl\" (UniqueName: \"kubernetes.io/projected/6ed10177-366b-4be3-b26a-f0c88841eb09-kube-api-access-ks7zl\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.592141 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ed10177-366b-4be3-b26a-f0c88841eb09","Type":"ContainerDied","Data":"d3083fc8db689be10b9d00a47018e0816a0143f0c99dcda294dfd71065d58997"} Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.592227 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.592240 4810 scope.go:117] "RemoveContainer" containerID="c8a6f55db2bed56e970121f4fd7ef18f7da3fde6130b6be3f31b65d0967d886b" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.642397 4810 scope.go:117] "RemoveContainer" containerID="56263de56fc274b57866a8ff801e7a382a9472f40897e29754ddd36ab39b25ba" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.658233 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.686406 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.699587 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:04 crc kubenswrapper[4810]: E0930 08:25:04.700146 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerName="nova-api-api" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.700173 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerName="nova-api-api" Sep 30 08:25:04 crc kubenswrapper[4810]: E0930 08:25:04.700192 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerName="nova-api-log" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.700202 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerName="nova-api-log" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.700516 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerName="nova-api-log" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.700585 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed10177-366b-4be3-b26a-f0c88841eb09" containerName="nova-api-api" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.702218 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.707569 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.707675 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.707837 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.708566 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.833442 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-config-data\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.833638 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d19b2ea-4e00-451a-86d6-4d503f57a10a-logs\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.833843 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.833915 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-public-tls-certs\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.834030 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.834188 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhbf5\" (UniqueName: \"kubernetes.io/projected/1d19b2ea-4e00-451a-86d6-4d503f57a10a-kube-api-access-dhbf5\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.936455 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.936544 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhbf5\" (UniqueName: \"kubernetes.io/projected/1d19b2ea-4e00-451a-86d6-4d503f57a10a-kube-api-access-dhbf5\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.936654 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-config-data\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.936698 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d19b2ea-4e00-451a-86d6-4d503f57a10a-logs\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.936762 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.936787 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-public-tls-certs\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.937745 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d19b2ea-4e00-451a-86d6-4d503f57a10a-logs\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.942099 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.942120 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-config-data\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.942586 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-public-tls-certs\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.943090 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:04 crc kubenswrapper[4810]: I0930 08:25:04.969772 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhbf5\" (UniqueName: \"kubernetes.io/projected/1d19b2ea-4e00-451a-86d6-4d503f57a10a-kube-api-access-dhbf5\") pod \"nova-api-0\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " pod="openstack/nova-api-0" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.029219 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.319214 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ed10177-366b-4be3-b26a-f0c88841eb09" path="/var/lib/kubelet/pods/6ed10177-366b-4be3-b26a-f0c88841eb09/volumes" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.557453 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.586741 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.617999 4810 generic.go:334] "Generic (PLEG): container finished" podID="825a301d-094e-4c79-8b68-2f038635df38" containerID="2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82" exitCode=0 Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.618040 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.618062 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"825a301d-094e-4c79-8b68-2f038635df38","Type":"ContainerDied","Data":"2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82"} Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.618130 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"825a301d-094e-4c79-8b68-2f038635df38","Type":"ContainerDied","Data":"c65d54f8ac0ec44dff3b04865c6a9a30210adabb4ca3803d5f7aae5bff4684c4"} Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.618165 4810 scope.go:117] "RemoveContainer" containerID="f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.619583 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d19b2ea-4e00-451a-86d6-4d503f57a10a","Type":"ContainerStarted","Data":"0da85c2382ff36a030670f3a0d9e86217e400b31a389a390098fcf53fceae496"} Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.652455 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-config-data\") pod \"825a301d-094e-4c79-8b68-2f038635df38\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.652586 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-run-httpd\") pod \"825a301d-094e-4c79-8b68-2f038635df38\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.652629 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hrbw\" (UniqueName: \"kubernetes.io/projected/825a301d-094e-4c79-8b68-2f038635df38-kube-api-access-7hrbw\") pod \"825a301d-094e-4c79-8b68-2f038635df38\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.652753 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-combined-ca-bundle\") pod \"825a301d-094e-4c79-8b68-2f038635df38\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.652793 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-ceilometer-tls-certs\") pod \"825a301d-094e-4c79-8b68-2f038635df38\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.653026 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "825a301d-094e-4c79-8b68-2f038635df38" (UID: "825a301d-094e-4c79-8b68-2f038635df38"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.653451 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-scripts\") pod \"825a301d-094e-4c79-8b68-2f038635df38\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.653495 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-sg-core-conf-yaml\") pod \"825a301d-094e-4c79-8b68-2f038635df38\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.653573 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-log-httpd\") pod \"825a301d-094e-4c79-8b68-2f038635df38\" (UID: \"825a301d-094e-4c79-8b68-2f038635df38\") " Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.654155 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.654567 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "825a301d-094e-4c79-8b68-2f038635df38" (UID: "825a301d-094e-4c79-8b68-2f038635df38"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.655900 4810 scope.go:117] "RemoveContainer" containerID="74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.658518 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/825a301d-094e-4c79-8b68-2f038635df38-kube-api-access-7hrbw" (OuterVolumeSpecName: "kube-api-access-7hrbw") pod "825a301d-094e-4c79-8b68-2f038635df38" (UID: "825a301d-094e-4c79-8b68-2f038635df38"). InnerVolumeSpecName "kube-api-access-7hrbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.661496 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-scripts" (OuterVolumeSpecName: "scripts") pod "825a301d-094e-4c79-8b68-2f038635df38" (UID: "825a301d-094e-4c79-8b68-2f038635df38"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.677230 4810 scope.go:117] "RemoveContainer" containerID="2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.681833 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "825a301d-094e-4c79-8b68-2f038635df38" (UID: "825a301d-094e-4c79-8b68-2f038635df38"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.704420 4810 scope.go:117] "RemoveContainer" containerID="24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.715795 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "825a301d-094e-4c79-8b68-2f038635df38" (UID: "825a301d-094e-4c79-8b68-2f038635df38"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.730317 4810 scope.go:117] "RemoveContainer" containerID="f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8" Sep 30 08:25:05 crc kubenswrapper[4810]: E0930 08:25:05.730908 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8\": container with ID starting with f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8 not found: ID does not exist" containerID="f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.730963 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8"} err="failed to get container status \"f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8\": rpc error: code = NotFound desc = could not find container \"f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8\": container with ID starting with f37cfcc5abb53b989bb33763ca620e71265ffc3dde25efe577f937ccf65614e8 not found: ID does not exist" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.730991 4810 scope.go:117] "RemoveContainer" containerID="74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f" Sep 30 08:25:05 crc kubenswrapper[4810]: E0930 08:25:05.731537 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f\": container with ID starting with 74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f not found: ID does not exist" containerID="74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.731563 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f"} err="failed to get container status \"74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f\": rpc error: code = NotFound desc = could not find container \"74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f\": container with ID starting with 74b3b63c54dde77e8d284233de1a881d5a6bfd6cd0f89151069d3caf792a424f not found: ID does not exist" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.731575 4810 scope.go:117] "RemoveContainer" containerID="2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82" Sep 30 08:25:05 crc kubenswrapper[4810]: E0930 08:25:05.732160 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82\": container with ID starting with 2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82 not found: ID does not exist" containerID="2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.732212 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82"} err="failed to get container status \"2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82\": rpc error: code = NotFound desc = could not find container \"2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82\": container with ID starting with 2011df2cba5b8ed3e8287dcf1ad119f5a9f6af88cc96c5259cdc4deb6ea65c82 not found: ID does not exist" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.732249 4810 scope.go:117] "RemoveContainer" containerID="24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e" Sep 30 08:25:05 crc kubenswrapper[4810]: E0930 08:25:05.732624 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e\": container with ID starting with 24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e not found: ID does not exist" containerID="24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.732655 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e"} err="failed to get container status \"24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e\": rpc error: code = NotFound desc = could not find container \"24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e\": container with ID starting with 24a3b4714b54fff6f587542bc897dc830e0fa80afbe8e5351f02ec2d9a51368e not found: ID does not exist" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.749344 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "825a301d-094e-4c79-8b68-2f038635df38" (UID: "825a301d-094e-4c79-8b68-2f038635df38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.756837 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.756881 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.756896 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.756909 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.756920 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/825a301d-094e-4c79-8b68-2f038635df38-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.756931 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hrbw\" (UniqueName: \"kubernetes.io/projected/825a301d-094e-4c79-8b68-2f038635df38-kube-api-access-7hrbw\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.768194 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-config-data" (OuterVolumeSpecName: "config-data") pod "825a301d-094e-4c79-8b68-2f038635df38" (UID: "825a301d-094e-4c79-8b68-2f038635df38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.859207 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825a301d-094e-4c79-8b68-2f038635df38-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.957394 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:25:05 crc kubenswrapper[4810]: I0930 08:25:05.987487 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.006395 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:25:06 crc kubenswrapper[4810]: E0930 08:25:06.006874 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="proxy-httpd" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.006892 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="proxy-httpd" Sep 30 08:25:06 crc kubenswrapper[4810]: E0930 08:25:06.006920 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="ceilometer-central-agent" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.006927 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="ceilometer-central-agent" Sep 30 08:25:06 crc kubenswrapper[4810]: E0930 08:25:06.006953 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="sg-core" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.006960 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="sg-core" Sep 30 08:25:06 crc kubenswrapper[4810]: E0930 08:25:06.006970 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="ceilometer-notification-agent" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.006976 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="ceilometer-notification-agent" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.007152 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="proxy-httpd" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.007168 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="sg-core" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.007182 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="ceilometer-notification-agent" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.007193 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="825a301d-094e-4c79-8b68-2f038635df38" containerName="ceilometer-central-agent" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.008984 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.013224 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.013570 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.013791 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.020231 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.168295 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.168671 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae076b6e-09cd-4911-89c3-e6edb61516ae-run-httpd\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.168789 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-config-data\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.168986 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.169389 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-scripts\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.169575 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.169780 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae076b6e-09cd-4911-89c3-e6edb61516ae-log-httpd\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.169882 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5bx2\" (UniqueName: \"kubernetes.io/projected/ae076b6e-09cd-4911-89c3-e6edb61516ae-kube-api-access-j5bx2\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.271797 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.271855 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-scripts\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.271936 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.271975 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae076b6e-09cd-4911-89c3-e6edb61516ae-log-httpd\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.272009 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5bx2\" (UniqueName: \"kubernetes.io/projected/ae076b6e-09cd-4911-89c3-e6edb61516ae-kube-api-access-j5bx2\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.272072 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.272099 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae076b6e-09cd-4911-89c3-e6edb61516ae-run-httpd\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.272137 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-config-data\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.272902 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae076b6e-09cd-4911-89c3-e6edb61516ae-run-httpd\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.273212 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae076b6e-09cd-4911-89c3-e6edb61516ae-log-httpd\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.276745 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.277023 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-config-data\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.277743 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-scripts\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.280330 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.291191 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae076b6e-09cd-4911-89c3-e6edb61516ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.295763 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5bx2\" (UniqueName: \"kubernetes.io/projected/ae076b6e-09cd-4911-89c3-e6edb61516ae-kube-api-access-j5bx2\") pod \"ceilometer-0\" (UID: \"ae076b6e-09cd-4911-89c3-e6edb61516ae\") " pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.377924 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.635319 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d19b2ea-4e00-451a-86d6-4d503f57a10a","Type":"ContainerStarted","Data":"fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3"} Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.635366 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d19b2ea-4e00-451a-86d6-4d503f57a10a","Type":"ContainerStarted","Data":"b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51"} Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.665616 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.665589077 podStartE2EDuration="2.665589077s" podCreationTimestamp="2025-09-30 08:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:25:06.65124522 +0000 UTC m=+1330.103444507" watchObservedRunningTime="2025-09-30 08:25:06.665589077 +0000 UTC m=+1330.117788344" Sep 30 08:25:06 crc kubenswrapper[4810]: I0930 08:25:06.866061 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 08:25:07 crc kubenswrapper[4810]: I0930 08:25:07.318129 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="825a301d-094e-4c79-8b68-2f038635df38" path="/var/lib/kubelet/pods/825a301d-094e-4c79-8b68-2f038635df38/volumes" Sep 30 08:25:07 crc kubenswrapper[4810]: I0930 08:25:07.650419 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae076b6e-09cd-4911-89c3-e6edb61516ae","Type":"ContainerStarted","Data":"6015e23cef418eedf5ec8fe780659602277ed282090e4945092fd82ed7676ee3"} Sep 30 08:25:07 crc kubenswrapper[4810]: I0930 08:25:07.650716 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae076b6e-09cd-4911-89c3-e6edb61516ae","Type":"ContainerStarted","Data":"79f4e75f7a7428fe9b0a010e6f729f02adb64c4cdb018cc3be488858870b69f7"} Sep 30 08:25:07 crc kubenswrapper[4810]: I0930 08:25:07.650736 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae076b6e-09cd-4911-89c3-e6edb61516ae","Type":"ContainerStarted","Data":"11eb124881e85b12885e41a739feb8a17e12204cb272f6c3386a462b7ffb1b4b"} Sep 30 08:25:08 crc kubenswrapper[4810]: I0930 08:25:08.664309 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae076b6e-09cd-4911-89c3-e6edb61516ae","Type":"ContainerStarted","Data":"d00cc4869f2a8ef2d650a8237b0e7827e016496c4481055fdcd4ea9d3bdca751"} Sep 30 08:25:08 crc kubenswrapper[4810]: I0930 08:25:08.952953 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:25:08 crc kubenswrapper[4810]: I0930 08:25:08.973342 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 08:25:08 crc kubenswrapper[4810]: I0930 08:25:08.973445 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 08:25:08 crc kubenswrapper[4810]: I0930 08:25:08.985325 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.690253 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.841964 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-pspj8"] Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.843600 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.845844 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.846075 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.857255 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pspj8"] Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.954307 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-scripts\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.954549 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz4w4\" (UniqueName: \"kubernetes.io/projected/a5df8e23-2303-4108-a55c-775633e3b4c9-kube-api-access-cz4w4\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.954663 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.954840 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-config-data\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.987485 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.221:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 08:25:09 crc kubenswrapper[4810]: I0930 08:25:09.987554 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.221:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.056256 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz4w4\" (UniqueName: \"kubernetes.io/projected/a5df8e23-2303-4108-a55c-775633e3b4c9-kube-api-access-cz4w4\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.056344 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.056425 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-config-data\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.056468 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-scripts\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.063594 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-config-data\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.071464 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.071875 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-scripts\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.074951 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz4w4\" (UniqueName: \"kubernetes.io/projected/a5df8e23-2303-4108-a55c-775633e3b4c9-kube-api-access-cz4w4\") pod \"nova-cell1-cell-mapping-pspj8\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.130975 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.164601 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.211464 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-569f96c6f5-dprqh"] Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.211721 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" podUID="8999f8b3-780a-4d07-a81a-e1aea00275c1" containerName="dnsmasq-dns" containerID="cri-o://8e139db0f1b5cfe766e8092b81405fed58e00152c429a7e5c26b2741746e4fd6" gracePeriod=10 Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.685671 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae076b6e-09cd-4911-89c3-e6edb61516ae","Type":"ContainerStarted","Data":"edc80c1c3755307b6425c74b73409c6241756c0e9088693c0eba8a19353da860"} Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.686306 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.699368 4810 generic.go:334] "Generic (PLEG): container finished" podID="8999f8b3-780a-4d07-a81a-e1aea00275c1" containerID="8e139db0f1b5cfe766e8092b81405fed58e00152c429a7e5c26b2741746e4fd6" exitCode=0 Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.700237 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" event={"ID":"8999f8b3-780a-4d07-a81a-e1aea00275c1","Type":"ContainerDied","Data":"8e139db0f1b5cfe766e8092b81405fed58e00152c429a7e5c26b2741746e4fd6"} Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.719029 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pspj8"] Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.725453 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.099371827 podStartE2EDuration="5.725433283s" podCreationTimestamp="2025-09-30 08:25:05 +0000 UTC" firstStartedPulling="2025-09-30 08:25:06.874498657 +0000 UTC m=+1330.326697924" lastFinishedPulling="2025-09-30 08:25:09.500560113 +0000 UTC m=+1332.952759380" observedRunningTime="2025-09-30 08:25:10.713049893 +0000 UTC m=+1334.165249160" watchObservedRunningTime="2025-09-30 08:25:10.725433283 +0000 UTC m=+1334.177632550" Sep 30 08:25:10 crc kubenswrapper[4810]: I0930 08:25:10.951350 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.076782 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-swift-storage-0\") pod \"8999f8b3-780a-4d07-a81a-e1aea00275c1\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.077180 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzhr6\" (UniqueName: \"kubernetes.io/projected/8999f8b3-780a-4d07-a81a-e1aea00275c1-kube-api-access-rzhr6\") pod \"8999f8b3-780a-4d07-a81a-e1aea00275c1\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.077202 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-nb\") pod \"8999f8b3-780a-4d07-a81a-e1aea00275c1\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.077246 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-config\") pod \"8999f8b3-780a-4d07-a81a-e1aea00275c1\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.077281 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-sb\") pod \"8999f8b3-780a-4d07-a81a-e1aea00275c1\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.077868 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-svc\") pod \"8999f8b3-780a-4d07-a81a-e1aea00275c1\" (UID: \"8999f8b3-780a-4d07-a81a-e1aea00275c1\") " Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.084441 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8999f8b3-780a-4d07-a81a-e1aea00275c1-kube-api-access-rzhr6" (OuterVolumeSpecName: "kube-api-access-rzhr6") pod "8999f8b3-780a-4d07-a81a-e1aea00275c1" (UID: "8999f8b3-780a-4d07-a81a-e1aea00275c1"). InnerVolumeSpecName "kube-api-access-rzhr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.147741 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8999f8b3-780a-4d07-a81a-e1aea00275c1" (UID: "8999f8b3-780a-4d07-a81a-e1aea00275c1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.149849 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8999f8b3-780a-4d07-a81a-e1aea00275c1" (UID: "8999f8b3-780a-4d07-a81a-e1aea00275c1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.159950 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-config" (OuterVolumeSpecName: "config") pod "8999f8b3-780a-4d07-a81a-e1aea00275c1" (UID: "8999f8b3-780a-4d07-a81a-e1aea00275c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.174032 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8999f8b3-780a-4d07-a81a-e1aea00275c1" (UID: "8999f8b3-780a-4d07-a81a-e1aea00275c1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.189472 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.189505 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzhr6\" (UniqueName: \"kubernetes.io/projected/8999f8b3-780a-4d07-a81a-e1aea00275c1-kube-api-access-rzhr6\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.189516 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.189525 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.189533 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.192749 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8999f8b3-780a-4d07-a81a-e1aea00275c1" (UID: "8999f8b3-780a-4d07-a81a-e1aea00275c1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.291644 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8999f8b3-780a-4d07-a81a-e1aea00275c1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.712017 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.712016 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569f96c6f5-dprqh" event={"ID":"8999f8b3-780a-4d07-a81a-e1aea00275c1","Type":"ContainerDied","Data":"38426afcb6e95c2e747919123d6314d9f682efebcc3a847064e147c407de1e26"} Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.712120 4810 scope.go:117] "RemoveContainer" containerID="8e139db0f1b5cfe766e8092b81405fed58e00152c429a7e5c26b2741746e4fd6" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.714565 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pspj8" event={"ID":"a5df8e23-2303-4108-a55c-775633e3b4c9","Type":"ContainerStarted","Data":"26150a9bcc58d9e492f465ced6fab9fc13397d1412a95f47629adc56a5c55140"} Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.714611 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pspj8" event={"ID":"a5df8e23-2303-4108-a55c-775633e3b4c9","Type":"ContainerStarted","Data":"1b5f7a014380db288630b560f62cf65baa5a3ea9eb6d24a9fc2c3062d2c5f0fb"} Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.733908 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-pspj8" podStartSLOduration=2.733892086 podStartE2EDuration="2.733892086s" podCreationTimestamp="2025-09-30 08:25:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:25:11.73229147 +0000 UTC m=+1335.184490737" watchObservedRunningTime="2025-09-30 08:25:11.733892086 +0000 UTC m=+1335.186091343" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.740017 4810 scope.go:117] "RemoveContainer" containerID="edec53c10496593782bc108fc5055a96e17b2b6b2cbd9935eafa0fbde695852a" Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.763654 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-569f96c6f5-dprqh"] Sep 30 08:25:11 crc kubenswrapper[4810]: I0930 08:25:11.774940 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-569f96c6f5-dprqh"] Sep 30 08:25:13 crc kubenswrapper[4810]: I0930 08:25:13.318417 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8999f8b3-780a-4d07-a81a-e1aea00275c1" path="/var/lib/kubelet/pods/8999f8b3-780a-4d07-a81a-e1aea00275c1/volumes" Sep 30 08:25:15 crc kubenswrapper[4810]: I0930 08:25:15.029456 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 08:25:15 crc kubenswrapper[4810]: I0930 08:25:15.029827 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 08:25:16 crc kubenswrapper[4810]: I0930 08:25:16.039593 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.223:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 08:25:16 crc kubenswrapper[4810]: I0930 08:25:16.050466 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.223:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 08:25:16 crc kubenswrapper[4810]: I0930 08:25:16.803094 4810 generic.go:334] "Generic (PLEG): container finished" podID="a5df8e23-2303-4108-a55c-775633e3b4c9" containerID="26150a9bcc58d9e492f465ced6fab9fc13397d1412a95f47629adc56a5c55140" exitCode=0 Sep 30 08:25:16 crc kubenswrapper[4810]: I0930 08:25:16.803153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pspj8" event={"ID":"a5df8e23-2303-4108-a55c-775633e3b4c9","Type":"ContainerDied","Data":"26150a9bcc58d9e492f465ced6fab9fc13397d1412a95f47629adc56a5c55140"} Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.281435 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.357142 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-combined-ca-bundle\") pod \"a5df8e23-2303-4108-a55c-775633e3b4c9\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.357227 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-config-data\") pod \"a5df8e23-2303-4108-a55c-775633e3b4c9\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.357330 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz4w4\" (UniqueName: \"kubernetes.io/projected/a5df8e23-2303-4108-a55c-775633e3b4c9-kube-api-access-cz4w4\") pod \"a5df8e23-2303-4108-a55c-775633e3b4c9\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.357507 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-scripts\") pod \"a5df8e23-2303-4108-a55c-775633e3b4c9\" (UID: \"a5df8e23-2303-4108-a55c-775633e3b4c9\") " Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.408321 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-config-data" (OuterVolumeSpecName: "config-data") pod "a5df8e23-2303-4108-a55c-775633e3b4c9" (UID: "a5df8e23-2303-4108-a55c-775633e3b4c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.409100 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5df8e23-2303-4108-a55c-775633e3b4c9" (UID: "a5df8e23-2303-4108-a55c-775633e3b4c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.408633 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5df8e23-2303-4108-a55c-775633e3b4c9-kube-api-access-cz4w4" (OuterVolumeSpecName: "kube-api-access-cz4w4") pod "a5df8e23-2303-4108-a55c-775633e3b4c9" (UID: "a5df8e23-2303-4108-a55c-775633e3b4c9"). InnerVolumeSpecName "kube-api-access-cz4w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.409374 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-scripts" (OuterVolumeSpecName: "scripts") pod "a5df8e23-2303-4108-a55c-775633e3b4c9" (UID: "a5df8e23-2303-4108-a55c-775633e3b4c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.460512 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.460713 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.460777 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz4w4\" (UniqueName: \"kubernetes.io/projected/a5df8e23-2303-4108-a55c-775633e3b4c9-kube-api-access-cz4w4\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.460845 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5df8e23-2303-4108-a55c-775633e3b4c9-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.838700 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pspj8" event={"ID":"a5df8e23-2303-4108-a55c-775633e3b4c9","Type":"ContainerDied","Data":"1b5f7a014380db288630b560f62cf65baa5a3ea9eb6d24a9fc2c3062d2c5f0fb"} Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.838768 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b5f7a014380db288630b560f62cf65baa5a3ea9eb6d24a9fc2c3062d2c5f0fb" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.838843 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pspj8" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.984184 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 08:25:18 crc kubenswrapper[4810]: I0930 08:25:18.990492 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.001114 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.072308 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.072715 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerName="nova-api-log" containerID="cri-o://b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51" gracePeriod=30 Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.072917 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerName="nova-api-api" containerID="cri-o://fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3" gracePeriod=30 Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.089442 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.093768 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.094299 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="60aa73ab-48bd-4955-a69e-ef1993fa420e" containerName="nova-scheduler-scheduler" containerID="cri-o://53c240cb16bb6f14f8acc07cc944df9a8fe30ec0f31adbd2c28781ac6153c810" gracePeriod=30 Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.853633 4810 generic.go:334] "Generic (PLEG): container finished" podID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerID="b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51" exitCode=143 Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.853909 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d19b2ea-4e00-451a-86d6-4d503f57a10a","Type":"ContainerDied","Data":"b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51"} Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.856574 4810 generic.go:334] "Generic (PLEG): container finished" podID="60aa73ab-48bd-4955-a69e-ef1993fa420e" containerID="53c240cb16bb6f14f8acc07cc944df9a8fe30ec0f31adbd2c28781ac6153c810" exitCode=0 Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.857970 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60aa73ab-48bd-4955-a69e-ef1993fa420e","Type":"ContainerDied","Data":"53c240cb16bb6f14f8acc07cc944df9a8fe30ec0f31adbd2c28781ac6153c810"} Sep 30 08:25:19 crc kubenswrapper[4810]: I0930 08:25:19.871817 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.154718 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.311934 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-config-data\") pod \"60aa73ab-48bd-4955-a69e-ef1993fa420e\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.312063 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4j9r\" (UniqueName: \"kubernetes.io/projected/60aa73ab-48bd-4955-a69e-ef1993fa420e-kube-api-access-h4j9r\") pod \"60aa73ab-48bd-4955-a69e-ef1993fa420e\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.312188 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-combined-ca-bundle\") pod \"60aa73ab-48bd-4955-a69e-ef1993fa420e\" (UID: \"60aa73ab-48bd-4955-a69e-ef1993fa420e\") " Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.317323 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60aa73ab-48bd-4955-a69e-ef1993fa420e-kube-api-access-h4j9r" (OuterVolumeSpecName: "kube-api-access-h4j9r") pod "60aa73ab-48bd-4955-a69e-ef1993fa420e" (UID: "60aa73ab-48bd-4955-a69e-ef1993fa420e"). InnerVolumeSpecName "kube-api-access-h4j9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.346160 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60aa73ab-48bd-4955-a69e-ef1993fa420e" (UID: "60aa73ab-48bd-4955-a69e-ef1993fa420e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.357064 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-config-data" (OuterVolumeSpecName: "config-data") pod "60aa73ab-48bd-4955-a69e-ef1993fa420e" (UID: "60aa73ab-48bd-4955-a69e-ef1993fa420e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.414562 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.414595 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4j9r\" (UniqueName: \"kubernetes.io/projected/60aa73ab-48bd-4955-a69e-ef1993fa420e-kube-api-access-h4j9r\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.414605 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60aa73ab-48bd-4955-a69e-ef1993fa420e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.446741 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.515591 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-internal-tls-certs\") pod \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.515671 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-config-data\") pod \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.515761 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhbf5\" (UniqueName: \"kubernetes.io/projected/1d19b2ea-4e00-451a-86d6-4d503f57a10a-kube-api-access-dhbf5\") pod \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.515889 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d19b2ea-4e00-451a-86d6-4d503f57a10a-logs\") pod \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.515987 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-combined-ca-bundle\") pod \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.516019 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-public-tls-certs\") pod \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\" (UID: \"1d19b2ea-4e00-451a-86d6-4d503f57a10a\") " Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.516627 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d19b2ea-4e00-451a-86d6-4d503f57a10a-logs" (OuterVolumeSpecName: "logs") pod "1d19b2ea-4e00-451a-86d6-4d503f57a10a" (UID: "1d19b2ea-4e00-451a-86d6-4d503f57a10a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.519427 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d19b2ea-4e00-451a-86d6-4d503f57a10a-kube-api-access-dhbf5" (OuterVolumeSpecName: "kube-api-access-dhbf5") pod "1d19b2ea-4e00-451a-86d6-4d503f57a10a" (UID: "1d19b2ea-4e00-451a-86d6-4d503f57a10a"). InnerVolumeSpecName "kube-api-access-dhbf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.547008 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-config-data" (OuterVolumeSpecName: "config-data") pod "1d19b2ea-4e00-451a-86d6-4d503f57a10a" (UID: "1d19b2ea-4e00-451a-86d6-4d503f57a10a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.568703 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d19b2ea-4e00-451a-86d6-4d503f57a10a" (UID: "1d19b2ea-4e00-451a-86d6-4d503f57a10a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.576837 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1d19b2ea-4e00-451a-86d6-4d503f57a10a" (UID: "1d19b2ea-4e00-451a-86d6-4d503f57a10a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.590821 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1d19b2ea-4e00-451a-86d6-4d503f57a10a" (UID: "1d19b2ea-4e00-451a-86d6-4d503f57a10a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.617919 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.617956 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.617967 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.617993 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d19b2ea-4e00-451a-86d6-4d503f57a10a-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.618004 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhbf5\" (UniqueName: \"kubernetes.io/projected/1d19b2ea-4e00-451a-86d6-4d503f57a10a-kube-api-access-dhbf5\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.618015 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d19b2ea-4e00-451a-86d6-4d503f57a10a-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.869094 4810 generic.go:334] "Generic (PLEG): container finished" podID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerID="fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3" exitCode=0 Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.869202 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d19b2ea-4e00-451a-86d6-4d503f57a10a","Type":"ContainerDied","Data":"fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3"} Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.869284 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d19b2ea-4e00-451a-86d6-4d503f57a10a","Type":"ContainerDied","Data":"0da85c2382ff36a030670f3a0d9e86217e400b31a389a390098fcf53fceae496"} Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.869309 4810 scope.go:117] "RemoveContainer" containerID="fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.870344 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.872394 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerName="nova-metadata-log" containerID="cri-o://e1004c44a710c6b1586c7ec58d8cb79dfbbe1bc7012bef84028057c07d6549a6" gracePeriod=30 Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.872565 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.874358 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60aa73ab-48bd-4955-a69e-ef1993fa420e","Type":"ContainerDied","Data":"a8c8904e8eede7adc0e24234724d5769a1896ea8bbbd7903e2f64ee32ca8f1a8"} Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.874383 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerName="nova-metadata-metadata" containerID="cri-o://b2e64049309fffd0cc128b6123caa07c1ccb1191f677273a1715030c3141fdbd" gracePeriod=30 Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.893874 4810 scope.go:117] "RemoveContainer" containerID="b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.916187 4810 scope.go:117] "RemoveContainer" containerID="fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3" Sep 30 08:25:20 crc kubenswrapper[4810]: E0930 08:25:20.916884 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3\": container with ID starting with fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3 not found: ID does not exist" containerID="fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.916922 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3"} err="failed to get container status \"fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3\": rpc error: code = NotFound desc = could not find container \"fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3\": container with ID starting with fa548fdd1ba428ccbaf2c7a8aa6fe20e18e28acc18b9efc77af416e82595dba3 not found: ID does not exist" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.916946 4810 scope.go:117] "RemoveContainer" containerID="b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51" Sep 30 08:25:20 crc kubenswrapper[4810]: E0930 08:25:20.917348 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51\": container with ID starting with b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51 not found: ID does not exist" containerID="b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.917405 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51"} err="failed to get container status \"b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51\": rpc error: code = NotFound desc = could not find container \"b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51\": container with ID starting with b8f6638b0104e363ef0d8026d97e1229a56bb2c296f58daa25338bfdfa9efa51 not found: ID does not exist" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.917439 4810 scope.go:117] "RemoveContainer" containerID="53c240cb16bb6f14f8acc07cc944df9a8fe30ec0f31adbd2c28781ac6153c810" Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.931343 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.942863 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.958183 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:20 crc kubenswrapper[4810]: I0930 08:25:20.968083 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991059 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:25:21 crc kubenswrapper[4810]: E0930 08:25:20.991453 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8999f8b3-780a-4d07-a81a-e1aea00275c1" containerName="dnsmasq-dns" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991467 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8999f8b3-780a-4d07-a81a-e1aea00275c1" containerName="dnsmasq-dns" Sep 30 08:25:21 crc kubenswrapper[4810]: E0930 08:25:20.991490 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60aa73ab-48bd-4955-a69e-ef1993fa420e" containerName="nova-scheduler-scheduler" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991496 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="60aa73ab-48bd-4955-a69e-ef1993fa420e" containerName="nova-scheduler-scheduler" Sep 30 08:25:21 crc kubenswrapper[4810]: E0930 08:25:20.991513 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5df8e23-2303-4108-a55c-775633e3b4c9" containerName="nova-manage" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991532 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5df8e23-2303-4108-a55c-775633e3b4c9" containerName="nova-manage" Sep 30 08:25:21 crc kubenswrapper[4810]: E0930 08:25:20.991548 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerName="nova-api-log" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991553 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerName="nova-api-log" Sep 30 08:25:21 crc kubenswrapper[4810]: E0930 08:25:20.991565 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerName="nova-api-api" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991571 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerName="nova-api-api" Sep 30 08:25:21 crc kubenswrapper[4810]: E0930 08:25:20.991581 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8999f8b3-780a-4d07-a81a-e1aea00275c1" containerName="init" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991587 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8999f8b3-780a-4d07-a81a-e1aea00275c1" containerName="init" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991751 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5df8e23-2303-4108-a55c-775633e3b4c9" containerName="nova-manage" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991768 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="60aa73ab-48bd-4955-a69e-ef1993fa420e" containerName="nova-scheduler-scheduler" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991777 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8999f8b3-780a-4d07-a81a-e1aea00275c1" containerName="dnsmasq-dns" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991788 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerName="nova-api-api" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.991797 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" containerName="nova-api-log" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.992349 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.992362 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.993533 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:20.994014 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.024975 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.025081 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.025173 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.025207 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.059473 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.137300 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a5417c-b645-430c-ada6-ae56ca4aa813-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"97a5417c-b645-430c-ada6-ae56ca4aa813\") " pod="openstack/nova-scheduler-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.137382 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a5417c-b645-430c-ada6-ae56ca4aa813-config-data\") pod \"nova-scheduler-0\" (UID: \"97a5417c-b645-430c-ada6-ae56ca4aa813\") " pod="openstack/nova-scheduler-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.137454 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.137518 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sth9\" (UniqueName: \"kubernetes.io/projected/97a5417c-b645-430c-ada6-ae56ca4aa813-kube-api-access-4sth9\") pod \"nova-scheduler-0\" (UID: \"97a5417c-b645-430c-ada6-ae56ca4aa813\") " pod="openstack/nova-scheduler-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.137586 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdw5k\" (UniqueName: \"kubernetes.io/projected/7e2cfe82-5249-4207-9178-f4ebb56a9f93-kube-api-access-xdw5k\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.137629 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.137753 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-config-data\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.137887 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-public-tls-certs\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.138042 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e2cfe82-5249-4207-9178-f4ebb56a9f93-logs\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.239702 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sth9\" (UniqueName: \"kubernetes.io/projected/97a5417c-b645-430c-ada6-ae56ca4aa813-kube-api-access-4sth9\") pod \"nova-scheduler-0\" (UID: \"97a5417c-b645-430c-ada6-ae56ca4aa813\") " pod="openstack/nova-scheduler-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.239777 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdw5k\" (UniqueName: \"kubernetes.io/projected/7e2cfe82-5249-4207-9178-f4ebb56a9f93-kube-api-access-xdw5k\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.239828 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.239866 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-config-data\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.239909 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-public-tls-certs\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.239948 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e2cfe82-5249-4207-9178-f4ebb56a9f93-logs\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.239999 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a5417c-b645-430c-ada6-ae56ca4aa813-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"97a5417c-b645-430c-ada6-ae56ca4aa813\") " pod="openstack/nova-scheduler-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.240035 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a5417c-b645-430c-ada6-ae56ca4aa813-config-data\") pod \"nova-scheduler-0\" (UID: \"97a5417c-b645-430c-ada6-ae56ca4aa813\") " pod="openstack/nova-scheduler-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.240065 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.241824 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e2cfe82-5249-4207-9178-f4ebb56a9f93-logs\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.245194 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.245553 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-config-data\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.245729 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a5417c-b645-430c-ada6-ae56ca4aa813-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"97a5417c-b645-430c-ada6-ae56ca4aa813\") " pod="openstack/nova-scheduler-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.245880 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-public-tls-certs\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.246511 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a5417c-b645-430c-ada6-ae56ca4aa813-config-data\") pod \"nova-scheduler-0\" (UID: \"97a5417c-b645-430c-ada6-ae56ca4aa813\") " pod="openstack/nova-scheduler-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.249087 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e2cfe82-5249-4207-9178-f4ebb56a9f93-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.261098 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sth9\" (UniqueName: \"kubernetes.io/projected/97a5417c-b645-430c-ada6-ae56ca4aa813-kube-api-access-4sth9\") pod \"nova-scheduler-0\" (UID: \"97a5417c-b645-430c-ada6-ae56ca4aa813\") " pod="openstack/nova-scheduler-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.261762 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdw5k\" (UniqueName: \"kubernetes.io/projected/7e2cfe82-5249-4207-9178-f4ebb56a9f93-kube-api-access-xdw5k\") pod \"nova-api-0\" (UID: \"7e2cfe82-5249-4207-9178-f4ebb56a9f93\") " pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.324462 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d19b2ea-4e00-451a-86d6-4d503f57a10a" path="/var/lib/kubelet/pods/1d19b2ea-4e00-451a-86d6-4d503f57a10a/volumes" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.325340 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60aa73ab-48bd-4955-a69e-ef1993fa420e" path="/var/lib/kubelet/pods/60aa73ab-48bd-4955-a69e-ef1993fa420e/volumes" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.351666 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.365105 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.893388 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.894202 4810 generic.go:334] "Generic (PLEG): container finished" podID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerID="e1004c44a710c6b1586c7ec58d8cb79dfbbe1bc7012bef84028057c07d6549a6" exitCode=143 Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.894331 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be553dcb-b399-458b-8e33-482a0cf48ba7","Type":"ContainerDied","Data":"e1004c44a710c6b1586c7ec58d8cb79dfbbe1bc7012bef84028057c07d6549a6"} Sep 30 08:25:21 crc kubenswrapper[4810]: I0930 08:25:21.952012 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 08:25:21 crc kubenswrapper[4810]: W0930 08:25:21.952882 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e2cfe82_5249_4207_9178_f4ebb56a9f93.slice/crio-256f7fdbd9b08a3ec31ff53c619b6be4d6fe75d73678aa06fc207d1be30e84ff WatchSource:0}: Error finding container 256f7fdbd9b08a3ec31ff53c619b6be4d6fe75d73678aa06fc207d1be30e84ff: Status 404 returned error can't find the container with id 256f7fdbd9b08a3ec31ff53c619b6be4d6fe75d73678aa06fc207d1be30e84ff Sep 30 08:25:22 crc kubenswrapper[4810]: I0930 08:25:22.918479 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"97a5417c-b645-430c-ada6-ae56ca4aa813","Type":"ContainerStarted","Data":"1edd9b1eeb3511a2ea25335105d00203abcb8ee6bb7d07e8ff7c57f3817f8057"} Sep 30 08:25:22 crc kubenswrapper[4810]: I0930 08:25:22.919377 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"97a5417c-b645-430c-ada6-ae56ca4aa813","Type":"ContainerStarted","Data":"aabf54a6282dbd9d6a79464c26d074e03321c9696f7a9c7323cad11989a64205"} Sep 30 08:25:22 crc kubenswrapper[4810]: I0930 08:25:22.922487 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7e2cfe82-5249-4207-9178-f4ebb56a9f93","Type":"ContainerStarted","Data":"b6422f38267b83e50918f52b1b8dc7e24b96d679cc5e6f9d8b3d78b1784aef05"} Sep 30 08:25:22 crc kubenswrapper[4810]: I0930 08:25:22.922530 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7e2cfe82-5249-4207-9178-f4ebb56a9f93","Type":"ContainerStarted","Data":"af2c096b0ac8748e37b2df8876878121481d3b054d164f13dbb75b600ef7e9e7"} Sep 30 08:25:22 crc kubenswrapper[4810]: I0930 08:25:22.922543 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7e2cfe82-5249-4207-9178-f4ebb56a9f93","Type":"ContainerStarted","Data":"256f7fdbd9b08a3ec31ff53c619b6be4d6fe75d73678aa06fc207d1be30e84ff"} Sep 30 08:25:22 crc kubenswrapper[4810]: I0930 08:25:22.925108 4810 generic.go:334] "Generic (PLEG): container finished" podID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerID="b2e64049309fffd0cc128b6123caa07c1ccb1191f677273a1715030c3141fdbd" exitCode=0 Sep 30 08:25:22 crc kubenswrapper[4810]: I0930 08:25:22.925233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be553dcb-b399-458b-8e33-482a0cf48ba7","Type":"ContainerDied","Data":"b2e64049309fffd0cc128b6123caa07c1ccb1191f677273a1715030c3141fdbd"} Sep 30 08:25:22 crc kubenswrapper[4810]: I0930 08:25:22.925371 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be553dcb-b399-458b-8e33-482a0cf48ba7","Type":"ContainerDied","Data":"d8ed2db606176813feca796e503e2711f07aed7aaf991d923f3ecbc4028efd28"} Sep 30 08:25:22 crc kubenswrapper[4810]: I0930 08:25:22.925403 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8ed2db606176813feca796e503e2711f07aed7aaf991d923f3ecbc4028efd28" Sep 30 08:25:22 crc kubenswrapper[4810]: I0930 08:25:22.953045 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.953022403 podStartE2EDuration="2.953022403s" podCreationTimestamp="2025-09-30 08:25:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:25:22.944460405 +0000 UTC m=+1346.396659682" watchObservedRunningTime="2025-09-30 08:25:22.953022403 +0000 UTC m=+1346.405221660" Sep 30 08:25:22 crc kubenswrapper[4810]: I0930 08:25:22.969645 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.969603142 podStartE2EDuration="2.969603142s" podCreationTimestamp="2025-09-30 08:25:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:25:22.962921629 +0000 UTC m=+1346.415120896" watchObservedRunningTime="2025-09-30 08:25:22.969603142 +0000 UTC m=+1346.421802469" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.016204 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.182829 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88n29\" (UniqueName: \"kubernetes.io/projected/be553dcb-b399-458b-8e33-482a0cf48ba7-kube-api-access-88n29\") pod \"be553dcb-b399-458b-8e33-482a0cf48ba7\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.182946 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-config-data\") pod \"be553dcb-b399-458b-8e33-482a0cf48ba7\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.183003 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be553dcb-b399-458b-8e33-482a0cf48ba7-logs\") pod \"be553dcb-b399-458b-8e33-482a0cf48ba7\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.183093 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-nova-metadata-tls-certs\") pod \"be553dcb-b399-458b-8e33-482a0cf48ba7\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.183143 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-combined-ca-bundle\") pod \"be553dcb-b399-458b-8e33-482a0cf48ba7\" (UID: \"be553dcb-b399-458b-8e33-482a0cf48ba7\") " Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.183585 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be553dcb-b399-458b-8e33-482a0cf48ba7-logs" (OuterVolumeSpecName: "logs") pod "be553dcb-b399-458b-8e33-482a0cf48ba7" (UID: "be553dcb-b399-458b-8e33-482a0cf48ba7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.189329 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be553dcb-b399-458b-8e33-482a0cf48ba7-kube-api-access-88n29" (OuterVolumeSpecName: "kube-api-access-88n29") pod "be553dcb-b399-458b-8e33-482a0cf48ba7" (UID: "be553dcb-b399-458b-8e33-482a0cf48ba7"). InnerVolumeSpecName "kube-api-access-88n29". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.218479 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-config-data" (OuterVolumeSpecName: "config-data") pod "be553dcb-b399-458b-8e33-482a0cf48ba7" (UID: "be553dcb-b399-458b-8e33-482a0cf48ba7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.223991 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be553dcb-b399-458b-8e33-482a0cf48ba7" (UID: "be553dcb-b399-458b-8e33-482a0cf48ba7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.248800 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "be553dcb-b399-458b-8e33-482a0cf48ba7" (UID: "be553dcb-b399-458b-8e33-482a0cf48ba7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.286192 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.286231 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be553dcb-b399-458b-8e33-482a0cf48ba7-logs\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.286243 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.286257 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be553dcb-b399-458b-8e33-482a0cf48ba7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.286285 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88n29\" (UniqueName: \"kubernetes.io/projected/be553dcb-b399-458b-8e33-482a0cf48ba7-kube-api-access-88n29\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.936852 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.970604 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:25:23 crc kubenswrapper[4810]: I0930 08:25:23.991214 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.008645 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:25:24 crc kubenswrapper[4810]: E0930 08:25:24.009201 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerName="nova-metadata-log" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.009221 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerName="nova-metadata-log" Sep 30 08:25:24 crc kubenswrapper[4810]: E0930 08:25:24.009257 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerName="nova-metadata-metadata" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.009335 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerName="nova-metadata-metadata" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.009590 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerName="nova-metadata-metadata" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.009619 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="be553dcb-b399-458b-8e33-482a0cf48ba7" containerName="nova-metadata-log" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.011035 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.013361 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.014688 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.036963 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.103993 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d733771-3677-4a2e-9eda-86144efccf8d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.104179 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8m7d\" (UniqueName: \"kubernetes.io/projected/3d733771-3677-4a2e-9eda-86144efccf8d-kube-api-access-f8m7d\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.104254 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d733771-3677-4a2e-9eda-86144efccf8d-config-data\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.104368 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d733771-3677-4a2e-9eda-86144efccf8d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.104400 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d733771-3677-4a2e-9eda-86144efccf8d-logs\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.206493 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d733771-3677-4a2e-9eda-86144efccf8d-config-data\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.206866 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d733771-3677-4a2e-9eda-86144efccf8d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.207100 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d733771-3677-4a2e-9eda-86144efccf8d-logs\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.207420 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d733771-3677-4a2e-9eda-86144efccf8d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.207580 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d733771-3677-4a2e-9eda-86144efccf8d-logs\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.207887 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8m7d\" (UniqueName: \"kubernetes.io/projected/3d733771-3677-4a2e-9eda-86144efccf8d-kube-api-access-f8m7d\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.212779 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d733771-3677-4a2e-9eda-86144efccf8d-config-data\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.212947 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d733771-3677-4a2e-9eda-86144efccf8d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.222397 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d733771-3677-4a2e-9eda-86144efccf8d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.223040 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8m7d\" (UniqueName: \"kubernetes.io/projected/3d733771-3677-4a2e-9eda-86144efccf8d-kube-api-access-f8m7d\") pod \"nova-metadata-0\" (UID: \"3d733771-3677-4a2e-9eda-86144efccf8d\") " pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.335482 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.824246 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 08:25:24 crc kubenswrapper[4810]: W0930 08:25:24.828058 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d733771_3677_4a2e_9eda_86144efccf8d.slice/crio-b11c918ba011a03478ca6e6a50c0138f164b050045ad488ec70cfb0502841b5c WatchSource:0}: Error finding container b11c918ba011a03478ca6e6a50c0138f164b050045ad488ec70cfb0502841b5c: Status 404 returned error can't find the container with id b11c918ba011a03478ca6e6a50c0138f164b050045ad488ec70cfb0502841b5c Sep 30 08:25:24 crc kubenswrapper[4810]: I0930 08:25:24.960608 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3d733771-3677-4a2e-9eda-86144efccf8d","Type":"ContainerStarted","Data":"b11c918ba011a03478ca6e6a50c0138f164b050045ad488ec70cfb0502841b5c"} Sep 30 08:25:25 crc kubenswrapper[4810]: I0930 08:25:25.319314 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be553dcb-b399-458b-8e33-482a0cf48ba7" path="/var/lib/kubelet/pods/be553dcb-b399-458b-8e33-482a0cf48ba7/volumes" Sep 30 08:25:25 crc kubenswrapper[4810]: I0930 08:25:25.977405 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3d733771-3677-4a2e-9eda-86144efccf8d","Type":"ContainerStarted","Data":"91ee709f4229b8c502bc734356eec4e49df0005707dee65caead085ff843a1aa"} Sep 30 08:25:25 crc kubenswrapper[4810]: I0930 08:25:25.977809 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3d733771-3677-4a2e-9eda-86144efccf8d","Type":"ContainerStarted","Data":"512f9bf02b7ad1ea795a2746db26ff116c548115cec6fcd382f2d98bdb3c5080"} Sep 30 08:25:26 crc kubenswrapper[4810]: I0930 08:25:26.015282 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.015235672 podStartE2EDuration="3.015235672s" podCreationTimestamp="2025-09-30 08:25:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:25:26.008545538 +0000 UTC m=+1349.460744845" watchObservedRunningTime="2025-09-30 08:25:26.015235672 +0000 UTC m=+1349.467434949" Sep 30 08:25:26 crc kubenswrapper[4810]: I0930 08:25:26.365374 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 08:25:29 crc kubenswrapper[4810]: I0930 08:25:29.338786 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 08:25:29 crc kubenswrapper[4810]: I0930 08:25:29.339176 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 08:25:31 crc kubenswrapper[4810]: I0930 08:25:31.352386 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 08:25:31 crc kubenswrapper[4810]: I0930 08:25:31.352731 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 08:25:31 crc kubenswrapper[4810]: I0930 08:25:31.365608 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 08:25:31 crc kubenswrapper[4810]: I0930 08:25:31.399514 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 08:25:32 crc kubenswrapper[4810]: I0930 08:25:32.099188 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 08:25:32 crc kubenswrapper[4810]: I0930 08:25:32.368534 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7e2cfe82-5249-4207-9178-f4ebb56a9f93" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.227:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 08:25:32 crc kubenswrapper[4810]: I0930 08:25:32.368552 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7e2cfe82-5249-4207-9178-f4ebb56a9f93" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.227:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 08:25:34 crc kubenswrapper[4810]: I0930 08:25:34.336186 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 08:25:34 crc kubenswrapper[4810]: I0930 08:25:34.337117 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 08:25:35 crc kubenswrapper[4810]: I0930 08:25:35.359589 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3d733771-3677-4a2e-9eda-86144efccf8d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.228:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 08:25:35 crc kubenswrapper[4810]: I0930 08:25:35.359589 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3d733771-3677-4a2e-9eda-86144efccf8d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.228:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 08:25:36 crc kubenswrapper[4810]: I0930 08:25:36.395388 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 30 08:25:41 crc kubenswrapper[4810]: I0930 08:25:41.370369 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 08:25:41 crc kubenswrapper[4810]: I0930 08:25:41.373250 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 08:25:41 crc kubenswrapper[4810]: I0930 08:25:41.373774 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 08:25:41 crc kubenswrapper[4810]: I0930 08:25:41.390542 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 08:25:42 crc kubenswrapper[4810]: I0930 08:25:42.174676 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 08:25:42 crc kubenswrapper[4810]: I0930 08:25:42.198822 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 08:25:44 crc kubenswrapper[4810]: I0930 08:25:44.345736 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 08:25:44 crc kubenswrapper[4810]: I0930 08:25:44.351024 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 08:25:44 crc kubenswrapper[4810]: I0930 08:25:44.355829 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 08:25:45 crc kubenswrapper[4810]: I0930 08:25:45.222999 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 08:25:53 crc kubenswrapper[4810]: I0930 08:25:53.438653 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 08:25:54 crc kubenswrapper[4810]: I0930 08:25:54.184198 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.431296 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-76hk9"] Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.444024 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.454387 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76hk9"] Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.529494 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-utilities\") pod \"community-operators-76hk9\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.529570 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92nl4\" (UniqueName: \"kubernetes.io/projected/3feb2e15-d0ba-43f6-a1a7-e77998996131-kube-api-access-92nl4\") pod \"community-operators-76hk9\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.529673 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-catalog-content\") pod \"community-operators-76hk9\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.631364 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-utilities\") pod \"community-operators-76hk9\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.631457 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92nl4\" (UniqueName: \"kubernetes.io/projected/3feb2e15-d0ba-43f6-a1a7-e77998996131-kube-api-access-92nl4\") pod \"community-operators-76hk9\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.631541 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-catalog-content\") pod \"community-operators-76hk9\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.631958 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-catalog-content\") pod \"community-operators-76hk9\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.631960 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-utilities\") pod \"community-operators-76hk9\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.668245 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92nl4\" (UniqueName: \"kubernetes.io/projected/3feb2e15-d0ba-43f6-a1a7-e77998996131-kube-api-access-92nl4\") pod \"community-operators-76hk9\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:25:55 crc kubenswrapper[4810]: I0930 08:25:55.772200 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:25:56 crc kubenswrapper[4810]: I0930 08:25:56.319577 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76hk9"] Sep 30 08:25:56 crc kubenswrapper[4810]: I0930 08:25:56.823221 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="5bb3b225-5221-417f-9914-737723ac52aa" containerName="rabbitmq" containerID="cri-o://cc31d0996668af86e34cd3c8dc8059da7acee4bb9d1c4051edba62b215337f5b" gracePeriod=604797 Sep 30 08:25:57 crc kubenswrapper[4810]: I0930 08:25:57.137869 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="5bb3b225-5221-417f-9914-737723ac52aa" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Sep 30 08:25:57 crc kubenswrapper[4810]: I0930 08:25:57.355370 4810 generic.go:334] "Generic (PLEG): container finished" podID="3feb2e15-d0ba-43f6-a1a7-e77998996131" containerID="a22869bf1ec5c18c562f0a9dfcf360146dc5e71ab439d27df1cdb22f03bc8465" exitCode=0 Sep 30 08:25:57 crc kubenswrapper[4810]: I0930 08:25:57.355416 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76hk9" event={"ID":"3feb2e15-d0ba-43f6-a1a7-e77998996131","Type":"ContainerDied","Data":"a22869bf1ec5c18c562f0a9dfcf360146dc5e71ab439d27df1cdb22f03bc8465"} Sep 30 08:25:57 crc kubenswrapper[4810]: I0930 08:25:57.355441 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76hk9" event={"ID":"3feb2e15-d0ba-43f6-a1a7-e77998996131","Type":"ContainerStarted","Data":"6882d7a0b44542d165aa261fbe7f6f99209ab6efffe31e9f023f5a7500d28f60"} Sep 30 08:25:57 crc kubenswrapper[4810]: I0930 08:25:57.413478 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="cf638129-9405-48c6-8458-143b297363fd" containerName="rabbitmq" containerID="cri-o://a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f" gracePeriod=604797 Sep 30 08:25:57 crc kubenswrapper[4810]: I0930 08:25:57.464095 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="cf638129-9405-48c6-8458-143b297363fd" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.374252 4810 generic.go:334] "Generic (PLEG): container finished" podID="5bb3b225-5221-417f-9914-737723ac52aa" containerID="cc31d0996668af86e34cd3c8dc8059da7acee4bb9d1c4051edba62b215337f5b" exitCode=0 Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.374357 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5bb3b225-5221-417f-9914-737723ac52aa","Type":"ContainerDied","Data":"cc31d0996668af86e34cd3c8dc8059da7acee4bb9d1c4051edba62b215337f5b"} Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.515157 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.710927 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"5bb3b225-5221-417f-9914-737723ac52aa\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.711346 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-server-conf\") pod \"5bb3b225-5221-417f-9914-737723ac52aa\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.711386 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htf9r\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-kube-api-access-htf9r\") pod \"5bb3b225-5221-417f-9914-737723ac52aa\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.711426 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-config-data\") pod \"5bb3b225-5221-417f-9914-737723ac52aa\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.711501 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-tls\") pod \"5bb3b225-5221-417f-9914-737723ac52aa\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.711580 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5bb3b225-5221-417f-9914-737723ac52aa-erlang-cookie-secret\") pod \"5bb3b225-5221-417f-9914-737723ac52aa\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.711614 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-erlang-cookie\") pod \"5bb3b225-5221-417f-9914-737723ac52aa\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.711664 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-plugins\") pod \"5bb3b225-5221-417f-9914-737723ac52aa\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.711737 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-confd\") pod \"5bb3b225-5221-417f-9914-737723ac52aa\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.711802 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5bb3b225-5221-417f-9914-737723ac52aa-pod-info\") pod \"5bb3b225-5221-417f-9914-737723ac52aa\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.711851 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-plugins-conf\") pod \"5bb3b225-5221-417f-9914-737723ac52aa\" (UID: \"5bb3b225-5221-417f-9914-737723ac52aa\") " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.715867 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5bb3b225-5221-417f-9914-737723ac52aa" (UID: "5bb3b225-5221-417f-9914-737723ac52aa"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.716356 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5bb3b225-5221-417f-9914-737723ac52aa" (UID: "5bb3b225-5221-417f-9914-737723ac52aa"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.716967 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5bb3b225-5221-417f-9914-737723ac52aa" (UID: "5bb3b225-5221-417f-9914-737723ac52aa"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.721984 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5bb3b225-5221-417f-9914-737723ac52aa-pod-info" (OuterVolumeSpecName: "pod-info") pod "5bb3b225-5221-417f-9914-737723ac52aa" (UID: "5bb3b225-5221-417f-9914-737723ac52aa"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.723837 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "5bb3b225-5221-417f-9914-737723ac52aa" (UID: "5bb3b225-5221-417f-9914-737723ac52aa"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.724225 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5bb3b225-5221-417f-9914-737723ac52aa" (UID: "5bb3b225-5221-417f-9914-737723ac52aa"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.724308 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-kube-api-access-htf9r" (OuterVolumeSpecName: "kube-api-access-htf9r") pod "5bb3b225-5221-417f-9914-737723ac52aa" (UID: "5bb3b225-5221-417f-9914-737723ac52aa"). InnerVolumeSpecName "kube-api-access-htf9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.725136 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bb3b225-5221-417f-9914-737723ac52aa-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5bb3b225-5221-417f-9914-737723ac52aa" (UID: "5bb3b225-5221-417f-9914-737723ac52aa"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.757097 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-config-data" (OuterVolumeSpecName: "config-data") pod "5bb3b225-5221-417f-9914-737723ac52aa" (UID: "5bb3b225-5221-417f-9914-737723ac52aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.817801 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.817838 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htf9r\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-kube-api-access-htf9r\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.817853 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.817866 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.817877 4810 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5bb3b225-5221-417f-9914-737723ac52aa-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.817891 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.817902 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.817912 4810 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5bb3b225-5221-417f-9914-737723ac52aa-pod-info\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.817922 4810 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.818448 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-server-conf" (OuterVolumeSpecName: "server-conf") pod "5bb3b225-5221-417f-9914-737723ac52aa" (UID: "5bb3b225-5221-417f-9914-737723ac52aa"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.848881 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.891548 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5bb3b225-5221-417f-9914-737723ac52aa" (UID: "5bb3b225-5221-417f-9914-737723ac52aa"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.919799 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5bb3b225-5221-417f-9914-737723ac52aa-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.919835 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.919846 4810 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5bb3b225-5221-417f-9914-737723ac52aa-server-conf\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:58 crc kubenswrapper[4810]: I0930 08:25:58.991708 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.122850 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-tls\") pod \"cf638129-9405-48c6-8458-143b297363fd\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.123147 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cf638129-9405-48c6-8458-143b297363fd-erlang-cookie-secret\") pod \"cf638129-9405-48c6-8458-143b297363fd\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.123286 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cf638129-9405-48c6-8458-143b297363fd-pod-info\") pod \"cf638129-9405-48c6-8458-143b297363fd\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.123440 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j54sx\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-kube-api-access-j54sx\") pod \"cf638129-9405-48c6-8458-143b297363fd\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.123541 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"cf638129-9405-48c6-8458-143b297363fd\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.123638 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-plugins-conf\") pod \"cf638129-9405-48c6-8458-143b297363fd\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.123733 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-erlang-cookie\") pod \"cf638129-9405-48c6-8458-143b297363fd\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.123827 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-server-conf\") pod \"cf638129-9405-48c6-8458-143b297363fd\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.123899 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-config-data\") pod \"cf638129-9405-48c6-8458-143b297363fd\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.123991 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-confd\") pod \"cf638129-9405-48c6-8458-143b297363fd\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.124073 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-plugins\") pod \"cf638129-9405-48c6-8458-143b297363fd\" (UID: \"cf638129-9405-48c6-8458-143b297363fd\") " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.130435 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "cf638129-9405-48c6-8458-143b297363fd" (UID: "cf638129-9405-48c6-8458-143b297363fd"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.131326 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "cf638129-9405-48c6-8458-143b297363fd" (UID: "cf638129-9405-48c6-8458-143b297363fd"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.131920 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "cf638129-9405-48c6-8458-143b297363fd" (UID: "cf638129-9405-48c6-8458-143b297363fd"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.133744 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/cf638129-9405-48c6-8458-143b297363fd-pod-info" (OuterVolumeSpecName: "pod-info") pod "cf638129-9405-48c6-8458-143b297363fd" (UID: "cf638129-9405-48c6-8458-143b297363fd"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.134112 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "cf638129-9405-48c6-8458-143b297363fd" (UID: "cf638129-9405-48c6-8458-143b297363fd"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.134218 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf638129-9405-48c6-8458-143b297363fd-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "cf638129-9405-48c6-8458-143b297363fd" (UID: "cf638129-9405-48c6-8458-143b297363fd"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.137846 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "cf638129-9405-48c6-8458-143b297363fd" (UID: "cf638129-9405-48c6-8458-143b297363fd"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.141826 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-kube-api-access-j54sx" (OuterVolumeSpecName: "kube-api-access-j54sx") pod "cf638129-9405-48c6-8458-143b297363fd" (UID: "cf638129-9405-48c6-8458-143b297363fd"). InnerVolumeSpecName "kube-api-access-j54sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.202660 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-config-data" (OuterVolumeSpecName: "config-data") pod "cf638129-9405-48c6-8458-143b297363fd" (UID: "cf638129-9405-48c6-8458-143b297363fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.226475 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.226518 4810 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cf638129-9405-48c6-8458-143b297363fd-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.226530 4810 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cf638129-9405-48c6-8458-143b297363fd-pod-info\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.226541 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j54sx\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-kube-api-access-j54sx\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.226578 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.226593 4810 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.226607 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.226618 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.226628 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.243455 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-server-conf" (OuterVolumeSpecName: "server-conf") pod "cf638129-9405-48c6-8458-143b297363fd" (UID: "cf638129-9405-48c6-8458-143b297363fd"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.262936 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.272094 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "cf638129-9405-48c6-8458-143b297363fd" (UID: "cf638129-9405-48c6-8458-143b297363fd"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.327947 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.327982 4810 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cf638129-9405-48c6-8458-143b297363fd-server-conf\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.327995 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cf638129-9405-48c6-8458-143b297363fd-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 30 08:25:59 crc kubenswrapper[4810]: E0930 08:25:59.393913 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bb3b225_5221_417f_9914_737723ac52aa.slice/crio-c3c7f0ab98f0a063109b6ecd0d4261b59e1118b8cc965b3ff513d4873de959b2\": RecentStats: unable to find data in memory cache]" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.403868 4810 generic.go:334] "Generic (PLEG): container finished" podID="3feb2e15-d0ba-43f6-a1a7-e77998996131" containerID="432a8aa3dc718544b98fccc330d4e236341b75948fb88cd3ef43130621fd8950" exitCode=0 Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.404128 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76hk9" event={"ID":"3feb2e15-d0ba-43f6-a1a7-e77998996131","Type":"ContainerDied","Data":"432a8aa3dc718544b98fccc330d4e236341b75948fb88cd3ef43130621fd8950"} Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.408576 4810 generic.go:334] "Generic (PLEG): container finished" podID="cf638129-9405-48c6-8458-143b297363fd" containerID="a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f" exitCode=0 Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.408661 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cf638129-9405-48c6-8458-143b297363fd","Type":"ContainerDied","Data":"a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f"} Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.408697 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cf638129-9405-48c6-8458-143b297363fd","Type":"ContainerDied","Data":"137f0c7fece1e342a6e0a26f2c3598092749e6b180d83bec687bf69776fa0d5e"} Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.408718 4810 scope.go:117] "RemoveContainer" containerID="a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.408895 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.418618 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5bb3b225-5221-417f-9914-737723ac52aa","Type":"ContainerDied","Data":"c3c7f0ab98f0a063109b6ecd0d4261b59e1118b8cc965b3ff513d4873de959b2"} Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.418877 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.443073 4810 scope.go:117] "RemoveContainer" containerID="8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.460321 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.491208 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.509549 4810 scope.go:117] "RemoveContainer" containerID="a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f" Sep 30 08:25:59 crc kubenswrapper[4810]: E0930 08:25:59.521496 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f\": container with ID starting with a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f not found: ID does not exist" containerID="a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.521562 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f"} err="failed to get container status \"a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f\": rpc error: code = NotFound desc = could not find container \"a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f\": container with ID starting with a03c2d8dd8272d548775ee7145656c200708e92a2c66827567c939070c83962f not found: ID does not exist" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.521600 4810 scope.go:117] "RemoveContainer" containerID="8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21" Sep 30 08:25:59 crc kubenswrapper[4810]: E0930 08:25:59.523406 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21\": container with ID starting with 8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21 not found: ID does not exist" containerID="8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.523450 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21"} err="failed to get container status \"8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21\": rpc error: code = NotFound desc = could not find container \"8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21\": container with ID starting with 8b5a6057d02678d70755c1fb3a5cc8dc05ab45927def3b8b3b366a655444ba21 not found: ID does not exist" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.523479 4810 scope.go:117] "RemoveContainer" containerID="cc31d0996668af86e34cd3c8dc8059da7acee4bb9d1c4051edba62b215337f5b" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.527551 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.542137 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.552740 4810 scope.go:117] "RemoveContainer" containerID="dba7b116bca716cd0a076f7032ee15fd1e500f765b290f96f285fbc493d3400f" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.555415 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 08:25:59 crc kubenswrapper[4810]: E0930 08:25:59.556078 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf638129-9405-48c6-8458-143b297363fd" containerName="setup-container" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.556182 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf638129-9405-48c6-8458-143b297363fd" containerName="setup-container" Sep 30 08:25:59 crc kubenswrapper[4810]: E0930 08:25:59.556304 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf638129-9405-48c6-8458-143b297363fd" containerName="rabbitmq" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.556400 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf638129-9405-48c6-8458-143b297363fd" containerName="rabbitmq" Sep 30 08:25:59 crc kubenswrapper[4810]: E0930 08:25:59.556484 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bb3b225-5221-417f-9914-737723ac52aa" containerName="setup-container" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.556556 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bb3b225-5221-417f-9914-737723ac52aa" containerName="setup-container" Sep 30 08:25:59 crc kubenswrapper[4810]: E0930 08:25:59.556662 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bb3b225-5221-417f-9914-737723ac52aa" containerName="rabbitmq" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.556756 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bb3b225-5221-417f-9914-737723ac52aa" containerName="rabbitmq" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.557114 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf638129-9405-48c6-8458-143b297363fd" containerName="rabbitmq" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.557215 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bb3b225-5221-417f-9914-737723ac52aa" containerName="rabbitmq" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.558956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.568351 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gx7b6" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.568396 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.568478 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.568487 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.568561 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.568648 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.570624 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.571796 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.576812 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.579968 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.580160 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.581929 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.583481 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.583811 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6gwgr" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.583987 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.585193 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.585315 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.591713 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737237 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2hpt\" (UniqueName: \"kubernetes.io/projected/d0344203-4b62-409c-af24-0e619dfad39a-kube-api-access-h2hpt\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737312 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737337 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737358 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737399 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737422 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d0344203-4b62-409c-af24-0e619dfad39a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737437 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b695439c-7bde-408e-b90a-0aa0ff0bc494-config-data\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737453 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d0344203-4b62-409c-af24-0e619dfad39a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737644 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d0344203-4b62-409c-af24-0e619dfad39a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737701 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6b9q\" (UniqueName: \"kubernetes.io/projected/b695439c-7bde-408e-b90a-0aa0ff0bc494-kube-api-access-v6b9q\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737790 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b695439c-7bde-408e-b90a-0aa0ff0bc494-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.737895 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.738027 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0344203-4b62-409c-af24-0e619dfad39a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.738064 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.738134 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b695439c-7bde-408e-b90a-0aa0ff0bc494-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.738161 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d0344203-4b62-409c-af24-0e619dfad39a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.738192 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.738220 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b695439c-7bde-408e-b90a-0aa0ff0bc494-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.738248 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.738342 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.738370 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.738393 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b695439c-7bde-408e-b90a-0aa0ff0bc494-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840163 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840205 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840247 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840282 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d0344203-4b62-409c-af24-0e619dfad39a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840303 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b695439c-7bde-408e-b90a-0aa0ff0bc494-config-data\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840322 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d0344203-4b62-409c-af24-0e619dfad39a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840364 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d0344203-4b62-409c-af24-0e619dfad39a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840382 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6b9q\" (UniqueName: \"kubernetes.io/projected/b695439c-7bde-408e-b90a-0aa0ff0bc494-kube-api-access-v6b9q\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840405 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b695439c-7bde-408e-b90a-0aa0ff0bc494-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840447 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840501 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0344203-4b62-409c-af24-0e619dfad39a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840528 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840565 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b695439c-7bde-408e-b90a-0aa0ff0bc494-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840588 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d0344203-4b62-409c-af24-0e619dfad39a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840611 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840631 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b695439c-7bde-408e-b90a-0aa0ff0bc494-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840652 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840690 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840714 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840736 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b695439c-7bde-408e-b90a-0aa0ff0bc494-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840778 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2hpt\" (UniqueName: \"kubernetes.io/projected/d0344203-4b62-409c-af24-0e619dfad39a-kube-api-access-h2hpt\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840809 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.840960 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.841208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.841451 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b695439c-7bde-408e-b90a-0aa0ff0bc494-config-data\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.841491 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.841711 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0344203-4b62-409c-af24-0e619dfad39a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.841908 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.842003 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d0344203-4b62-409c-af24-0e619dfad39a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.842308 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d0344203-4b62-409c-af24-0e619dfad39a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.842370 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.842420 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.842748 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b695439c-7bde-408e-b90a-0aa0ff0bc494-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.842830 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b695439c-7bde-408e-b90a-0aa0ff0bc494-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.845833 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d0344203-4b62-409c-af24-0e619dfad39a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.846696 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b695439c-7bde-408e-b90a-0aa0ff0bc494-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.846732 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d0344203-4b62-409c-af24-0e619dfad39a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.847081 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.853970 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b695439c-7bde-408e-b90a-0aa0ff0bc494-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.854435 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b695439c-7bde-408e-b90a-0aa0ff0bc494-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.857296 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.861108 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6b9q\" (UniqueName: \"kubernetes.io/projected/b695439c-7bde-408e-b90a-0aa0ff0bc494-kube-api-access-v6b9q\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.861147 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d0344203-4b62-409c-af24-0e619dfad39a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.866598 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2hpt\" (UniqueName: \"kubernetes.io/projected/d0344203-4b62-409c-af24-0e619dfad39a-kube-api-access-h2hpt\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.901049 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0344203-4b62-409c-af24-0e619dfad39a\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:25:59 crc kubenswrapper[4810]: I0930 08:25:59.952412 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"b695439c-7bde-408e-b90a-0aa0ff0bc494\") " pod="openstack/rabbitmq-server-0" Sep 30 08:26:00 crc kubenswrapper[4810]: I0930 08:26:00.185186 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:26:00 crc kubenswrapper[4810]: I0930 08:26:00.215594 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 08:26:00 crc kubenswrapper[4810]: I0930 08:26:00.435022 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76hk9" event={"ID":"3feb2e15-d0ba-43f6-a1a7-e77998996131","Type":"ContainerStarted","Data":"42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b"} Sep 30 08:26:00 crc kubenswrapper[4810]: I0930 08:26:00.459603 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-76hk9" podStartSLOduration=2.86555214 podStartE2EDuration="5.459582422s" podCreationTimestamp="2025-09-30 08:25:55 +0000 UTC" firstStartedPulling="2025-09-30 08:25:57.35863492 +0000 UTC m=+1380.810834187" lastFinishedPulling="2025-09-30 08:25:59.952665192 +0000 UTC m=+1383.404864469" observedRunningTime="2025-09-30 08:26:00.457921434 +0000 UTC m=+1383.910120701" watchObservedRunningTime="2025-09-30 08:26:00.459582422 +0000 UTC m=+1383.911781689" Sep 30 08:26:00 crc kubenswrapper[4810]: I0930 08:26:00.665731 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 08:26:00 crc kubenswrapper[4810]: W0930 08:26:00.673301 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0344203_4b62_409c_af24_0e619dfad39a.slice/crio-310dc0d8ee0884a1737afbede21f44a9d8f01295ca90e13a7a675452e62ec5ba WatchSource:0}: Error finding container 310dc0d8ee0884a1737afbede21f44a9d8f01295ca90e13a7a675452e62ec5ba: Status 404 returned error can't find the container with id 310dc0d8ee0884a1737afbede21f44a9d8f01295ca90e13a7a675452e62ec5ba Sep 30 08:26:00 crc kubenswrapper[4810]: I0930 08:26:00.740226 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 08:26:00 crc kubenswrapper[4810]: W0930 08:26:00.753733 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb695439c_7bde_408e_b90a_0aa0ff0bc494.slice/crio-f6de867d275331cc59571a1661fd1eb30ca39b8d6b494e2770fe96f2da3c40cc WatchSource:0}: Error finding container f6de867d275331cc59571a1661fd1eb30ca39b8d6b494e2770fe96f2da3c40cc: Status 404 returned error can't find the container with id f6de867d275331cc59571a1661fd1eb30ca39b8d6b494e2770fe96f2da3c40cc Sep 30 08:26:01 crc kubenswrapper[4810]: I0930 08:26:01.320206 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bb3b225-5221-417f-9914-737723ac52aa" path="/var/lib/kubelet/pods/5bb3b225-5221-417f-9914-737723ac52aa/volumes" Sep 30 08:26:01 crc kubenswrapper[4810]: I0930 08:26:01.322321 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf638129-9405-48c6-8458-143b297363fd" path="/var/lib/kubelet/pods/cf638129-9405-48c6-8458-143b297363fd/volumes" Sep 30 08:26:01 crc kubenswrapper[4810]: I0930 08:26:01.451387 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b695439c-7bde-408e-b90a-0aa0ff0bc494","Type":"ContainerStarted","Data":"94380eeeebea55ed64cd5a48169987341384a32a723ebc853581689bbfa19b46"} Sep 30 08:26:01 crc kubenswrapper[4810]: I0930 08:26:01.451452 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b695439c-7bde-408e-b90a-0aa0ff0bc494","Type":"ContainerStarted","Data":"f6de867d275331cc59571a1661fd1eb30ca39b8d6b494e2770fe96f2da3c40cc"} Sep 30 08:26:01 crc kubenswrapper[4810]: I0930 08:26:01.454771 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d0344203-4b62-409c-af24-0e619dfad39a","Type":"ContainerStarted","Data":"bc619bbf4dff258f9e44808baef181ad87cdafbb25c295a7d6dd54b8bee3f2a3"} Sep 30 08:26:01 crc kubenswrapper[4810]: I0930 08:26:01.454819 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d0344203-4b62-409c-af24-0e619dfad39a","Type":"ContainerStarted","Data":"310dc0d8ee0884a1737afbede21f44a9d8f01295ca90e13a7a675452e62ec5ba"} Sep 30 08:26:05 crc kubenswrapper[4810]: I0930 08:26:05.773484 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:26:05 crc kubenswrapper[4810]: I0930 08:26:05.775617 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:26:05 crc kubenswrapper[4810]: I0930 08:26:05.860649 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:26:06 crc kubenswrapper[4810]: I0930 08:26:06.584317 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:26:06 crc kubenswrapper[4810]: I0930 08:26:06.647222 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76hk9"] Sep 30 08:26:08 crc kubenswrapper[4810]: I0930 08:26:08.541398 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-76hk9" podUID="3feb2e15-d0ba-43f6-a1a7-e77998996131" containerName="registry-server" containerID="cri-o://42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b" gracePeriod=2 Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.013877 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.141998 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-catalog-content\") pod \"3feb2e15-d0ba-43f6-a1a7-e77998996131\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.142155 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-utilities\") pod \"3feb2e15-d0ba-43f6-a1a7-e77998996131\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.142242 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92nl4\" (UniqueName: \"kubernetes.io/projected/3feb2e15-d0ba-43f6-a1a7-e77998996131-kube-api-access-92nl4\") pod \"3feb2e15-d0ba-43f6-a1a7-e77998996131\" (UID: \"3feb2e15-d0ba-43f6-a1a7-e77998996131\") " Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.143155 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-utilities" (OuterVolumeSpecName: "utilities") pod "3feb2e15-d0ba-43f6-a1a7-e77998996131" (UID: "3feb2e15-d0ba-43f6-a1a7-e77998996131"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.148783 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3feb2e15-d0ba-43f6-a1a7-e77998996131-kube-api-access-92nl4" (OuterVolumeSpecName: "kube-api-access-92nl4") pod "3feb2e15-d0ba-43f6-a1a7-e77998996131" (UID: "3feb2e15-d0ba-43f6-a1a7-e77998996131"). InnerVolumeSpecName "kube-api-access-92nl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.188749 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3feb2e15-d0ba-43f6-a1a7-e77998996131" (UID: "3feb2e15-d0ba-43f6-a1a7-e77998996131"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.246081 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.246124 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92nl4\" (UniqueName: \"kubernetes.io/projected/3feb2e15-d0ba-43f6-a1a7-e77998996131-kube-api-access-92nl4\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.246168 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feb2e15-d0ba-43f6-a1a7-e77998996131-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.322704 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cbfdf555-8fpkr"] Sep 30 08:26:09 crc kubenswrapper[4810]: E0930 08:26:09.323115 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feb2e15-d0ba-43f6-a1a7-e77998996131" containerName="extract-content" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.323185 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feb2e15-d0ba-43f6-a1a7-e77998996131" containerName="extract-content" Sep 30 08:26:09 crc kubenswrapper[4810]: E0930 08:26:09.323207 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feb2e15-d0ba-43f6-a1a7-e77998996131" containerName="extract-utilities" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.323215 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feb2e15-d0ba-43f6-a1a7-e77998996131" containerName="extract-utilities" Sep 30 08:26:09 crc kubenswrapper[4810]: E0930 08:26:09.323236 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feb2e15-d0ba-43f6-a1a7-e77998996131" containerName="registry-server" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.323335 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feb2e15-d0ba-43f6-a1a7-e77998996131" containerName="registry-server" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.323582 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3feb2e15-d0ba-43f6-a1a7-e77998996131" containerName="registry-server" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.324761 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cbfdf555-8fpkr"] Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.324870 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.327186 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.448545 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-svc\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.448630 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-openstack-edpm-ipam\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.448734 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pxh4\" (UniqueName: \"kubernetes.io/projected/6a51583a-0d9a-4f85-8768-8ff462be2485-kube-api-access-9pxh4\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.448834 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-sb\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.448868 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-swift-storage-0\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.448905 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-config\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.448955 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-nb\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.551352 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-svc\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.551437 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-openstack-edpm-ipam\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.551553 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pxh4\" (UniqueName: \"kubernetes.io/projected/6a51583a-0d9a-4f85-8768-8ff462be2485-kube-api-access-9pxh4\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.551692 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-sb\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.551746 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-swift-storage-0\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.551817 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-config\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.551898 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-nb\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.553811 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-openstack-edpm-ipam\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.554199 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-swift-storage-0\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.554464 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-nb\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.554499 4810 generic.go:334] "Generic (PLEG): container finished" podID="3feb2e15-d0ba-43f6-a1a7-e77998996131" containerID="42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b" exitCode=0 Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.554536 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76hk9" event={"ID":"3feb2e15-d0ba-43f6-a1a7-e77998996131","Type":"ContainerDied","Data":"42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b"} Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.554569 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76hk9" event={"ID":"3feb2e15-d0ba-43f6-a1a7-e77998996131","Type":"ContainerDied","Data":"6882d7a0b44542d165aa261fbe7f6f99209ab6efffe31e9f023f5a7500d28f60"} Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.554596 4810 scope.go:117] "RemoveContainer" containerID="42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.554665 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76hk9" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.555031 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-sb\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.555118 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-config\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.556115 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-svc\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.581489 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pxh4\" (UniqueName: \"kubernetes.io/projected/6a51583a-0d9a-4f85-8768-8ff462be2485-kube-api-access-9pxh4\") pod \"dnsmasq-dns-59cbfdf555-8fpkr\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.585122 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76hk9"] Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.588758 4810 scope.go:117] "RemoveContainer" containerID="432a8aa3dc718544b98fccc330d4e236341b75948fb88cd3ef43130621fd8950" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.594165 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-76hk9"] Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.651984 4810 scope.go:117] "RemoveContainer" containerID="a22869bf1ec5c18c562f0a9dfcf360146dc5e71ab439d27df1cdb22f03bc8465" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.656515 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.795990 4810 scope.go:117] "RemoveContainer" containerID="42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b" Sep 30 08:26:09 crc kubenswrapper[4810]: E0930 08:26:09.797677 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b\": container with ID starting with 42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b not found: ID does not exist" containerID="42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.797709 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b"} err="failed to get container status \"42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b\": rpc error: code = NotFound desc = could not find container \"42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b\": container with ID starting with 42b6bef4d3b298b27b3cfc16bb4b976217bc25f9b275583cbc6b97bde282b71b not found: ID does not exist" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.797731 4810 scope.go:117] "RemoveContainer" containerID="432a8aa3dc718544b98fccc330d4e236341b75948fb88cd3ef43130621fd8950" Sep 30 08:26:09 crc kubenswrapper[4810]: E0930 08:26:09.798040 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"432a8aa3dc718544b98fccc330d4e236341b75948fb88cd3ef43130621fd8950\": container with ID starting with 432a8aa3dc718544b98fccc330d4e236341b75948fb88cd3ef43130621fd8950 not found: ID does not exist" containerID="432a8aa3dc718544b98fccc330d4e236341b75948fb88cd3ef43130621fd8950" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.798095 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"432a8aa3dc718544b98fccc330d4e236341b75948fb88cd3ef43130621fd8950"} err="failed to get container status \"432a8aa3dc718544b98fccc330d4e236341b75948fb88cd3ef43130621fd8950\": rpc error: code = NotFound desc = could not find container \"432a8aa3dc718544b98fccc330d4e236341b75948fb88cd3ef43130621fd8950\": container with ID starting with 432a8aa3dc718544b98fccc330d4e236341b75948fb88cd3ef43130621fd8950 not found: ID does not exist" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.798125 4810 scope.go:117] "RemoveContainer" containerID="a22869bf1ec5c18c562f0a9dfcf360146dc5e71ab439d27df1cdb22f03bc8465" Sep 30 08:26:09 crc kubenswrapper[4810]: E0930 08:26:09.798460 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a22869bf1ec5c18c562f0a9dfcf360146dc5e71ab439d27df1cdb22f03bc8465\": container with ID starting with a22869bf1ec5c18c562f0a9dfcf360146dc5e71ab439d27df1cdb22f03bc8465 not found: ID does not exist" containerID="a22869bf1ec5c18c562f0a9dfcf360146dc5e71ab439d27df1cdb22f03bc8465" Sep 30 08:26:09 crc kubenswrapper[4810]: I0930 08:26:09.798484 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a22869bf1ec5c18c562f0a9dfcf360146dc5e71ab439d27df1cdb22f03bc8465"} err="failed to get container status \"a22869bf1ec5c18c562f0a9dfcf360146dc5e71ab439d27df1cdb22f03bc8465\": rpc error: code = NotFound desc = could not find container \"a22869bf1ec5c18c562f0a9dfcf360146dc5e71ab439d27df1cdb22f03bc8465\": container with ID starting with a22869bf1ec5c18c562f0a9dfcf360146dc5e71ab439d27df1cdb22f03bc8465 not found: ID does not exist" Sep 30 08:26:10 crc kubenswrapper[4810]: I0930 08:26:10.143461 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cbfdf555-8fpkr"] Sep 30 08:26:10 crc kubenswrapper[4810]: W0930 08:26:10.146824 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a51583a_0d9a_4f85_8768_8ff462be2485.slice/crio-73de7f84baf4f9d191c53d9612150ededfae67df86fa51a678168f7147543c40 WatchSource:0}: Error finding container 73de7f84baf4f9d191c53d9612150ededfae67df86fa51a678168f7147543c40: Status 404 returned error can't find the container with id 73de7f84baf4f9d191c53d9612150ededfae67df86fa51a678168f7147543c40 Sep 30 08:26:10 crc kubenswrapper[4810]: I0930 08:26:10.566259 4810 generic.go:334] "Generic (PLEG): container finished" podID="6a51583a-0d9a-4f85-8768-8ff462be2485" containerID="ef0de82656893412cdb8c0deb6734b4b0c1bf728448dcd0e3a89874ae3bba4c7" exitCode=0 Sep 30 08:26:10 crc kubenswrapper[4810]: I0930 08:26:10.566379 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" event={"ID":"6a51583a-0d9a-4f85-8768-8ff462be2485","Type":"ContainerDied","Data":"ef0de82656893412cdb8c0deb6734b4b0c1bf728448dcd0e3a89874ae3bba4c7"} Sep 30 08:26:10 crc kubenswrapper[4810]: I0930 08:26:10.566727 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" event={"ID":"6a51583a-0d9a-4f85-8768-8ff462be2485","Type":"ContainerStarted","Data":"73de7f84baf4f9d191c53d9612150ededfae67df86fa51a678168f7147543c40"} Sep 30 08:26:11 crc kubenswrapper[4810]: I0930 08:26:11.317865 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3feb2e15-d0ba-43f6-a1a7-e77998996131" path="/var/lib/kubelet/pods/3feb2e15-d0ba-43f6-a1a7-e77998996131/volumes" Sep 30 08:26:11 crc kubenswrapper[4810]: I0930 08:26:11.585624 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" event={"ID":"6a51583a-0d9a-4f85-8768-8ff462be2485","Type":"ContainerStarted","Data":"1fccb98ff7343b898c2ff0e27bbf9f156471d551d329b7938b1a77fd54308054"} Sep 30 08:26:11 crc kubenswrapper[4810]: I0930 08:26:11.585821 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:11 crc kubenswrapper[4810]: I0930 08:26:11.623115 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" podStartSLOduration=2.623094977 podStartE2EDuration="2.623094977s" podCreationTimestamp="2025-09-30 08:26:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:26:11.618539865 +0000 UTC m=+1395.070739172" watchObservedRunningTime="2025-09-30 08:26:11.623094977 +0000 UTC m=+1395.075294244" Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.440544 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2snhw"] Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.445734 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.457362 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2snhw"] Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.542370 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4mpp\" (UniqueName: \"kubernetes.io/projected/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-kube-api-access-c4mpp\") pod \"certified-operators-2snhw\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.542639 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-utilities\") pod \"certified-operators-2snhw\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.542684 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-catalog-content\") pod \"certified-operators-2snhw\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.645181 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-utilities\") pod \"certified-operators-2snhw\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.645251 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-catalog-content\") pod \"certified-operators-2snhw\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.645325 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4mpp\" (UniqueName: \"kubernetes.io/projected/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-kube-api-access-c4mpp\") pod \"certified-operators-2snhw\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.646165 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-utilities\") pod \"certified-operators-2snhw\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.647172 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-catalog-content\") pod \"certified-operators-2snhw\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.668850 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4mpp\" (UniqueName: \"kubernetes.io/projected/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-kube-api-access-c4mpp\") pod \"certified-operators-2snhw\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:13 crc kubenswrapper[4810]: I0930 08:26:13.782802 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:14 crc kubenswrapper[4810]: I0930 08:26:14.294846 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2snhw"] Sep 30 08:26:14 crc kubenswrapper[4810]: I0930 08:26:14.626116 4810 generic.go:334] "Generic (PLEG): container finished" podID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" containerID="6f0a175dabbc3d169b283d81207006f57217968cf6a3cdcd6de59518d2b722c7" exitCode=0 Sep 30 08:26:14 crc kubenswrapper[4810]: I0930 08:26:14.626193 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2snhw" event={"ID":"e10b0680-a0ad-45dd-8abf-10f91f4aba4c","Type":"ContainerDied","Data":"6f0a175dabbc3d169b283d81207006f57217968cf6a3cdcd6de59518d2b722c7"} Sep 30 08:26:14 crc kubenswrapper[4810]: I0930 08:26:14.626240 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2snhw" event={"ID":"e10b0680-a0ad-45dd-8abf-10f91f4aba4c","Type":"ContainerStarted","Data":"0e6e65e5c17e22416d3a4a85ac7e27fe1eb18e23c72440dc5fc72c00dabcb4e5"} Sep 30 08:26:15 crc kubenswrapper[4810]: I0930 08:26:15.638923 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2snhw" event={"ID":"e10b0680-a0ad-45dd-8abf-10f91f4aba4c","Type":"ContainerStarted","Data":"8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f"} Sep 30 08:26:16 crc kubenswrapper[4810]: I0930 08:26:16.650648 4810 generic.go:334] "Generic (PLEG): container finished" podID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" containerID="8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f" exitCode=0 Sep 30 08:26:16 crc kubenswrapper[4810]: I0930 08:26:16.651323 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2snhw" event={"ID":"e10b0680-a0ad-45dd-8abf-10f91f4aba4c","Type":"ContainerDied","Data":"8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f"} Sep 30 08:26:17 crc kubenswrapper[4810]: I0930 08:26:17.661322 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2snhw" event={"ID":"e10b0680-a0ad-45dd-8abf-10f91f4aba4c","Type":"ContainerStarted","Data":"68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c"} Sep 30 08:26:17 crc kubenswrapper[4810]: I0930 08:26:17.686686 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2snhw" podStartSLOduration=2.158372179 podStartE2EDuration="4.686665029s" podCreationTimestamp="2025-09-30 08:26:13 +0000 UTC" firstStartedPulling="2025-09-30 08:26:14.628415901 +0000 UTC m=+1398.080615168" lastFinishedPulling="2025-09-30 08:26:17.156708751 +0000 UTC m=+1400.608908018" observedRunningTime="2025-09-30 08:26:17.682620601 +0000 UTC m=+1401.134819958" watchObservedRunningTime="2025-09-30 08:26:17.686665029 +0000 UTC m=+1401.138864306" Sep 30 08:26:19 crc kubenswrapper[4810]: I0930 08:26:19.659696 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:19 crc kubenswrapper[4810]: I0930 08:26:19.758382 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb9f56f5-cg226"] Sep 30 08:26:19 crc kubenswrapper[4810]: I0930 08:26:19.758809 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" podUID="10c33784-e317-418b-bdf5-ea60acd29d98" containerName="dnsmasq-dns" containerID="cri-o://24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee" gracePeriod=10 Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.003550 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-585f78d44c-nlbbz"] Sep 30 08:26:20 crc kubenswrapper[4810]: E0930 08:26:20.014514 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10c33784_e317_418b_bdf5_ea60acd29d98.slice/crio-24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10c33784_e317_418b_bdf5_ea60acd29d98.slice/crio-conmon-24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee.scope\": RecentStats: unable to find data in memory cache]" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.020554 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-585f78d44c-nlbbz"] Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.020678 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.093639 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-dns-svc\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.093722 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-config\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.093750 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc7mz\" (UniqueName: \"kubernetes.io/projected/cf951386-0eef-4a7f-9247-b854b9a39d7b-kube-api-access-pc7mz\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.093775 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-dns-swift-storage-0\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.093799 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-ovsdbserver-nb\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.093895 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-ovsdbserver-sb\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.093954 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-openstack-edpm-ipam\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.195150 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-ovsdbserver-sb\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.195226 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-openstack-edpm-ipam\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.195290 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-dns-svc\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.195337 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-config\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.195356 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc7mz\" (UniqueName: \"kubernetes.io/projected/cf951386-0eef-4a7f-9247-b854b9a39d7b-kube-api-access-pc7mz\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.195373 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-dns-swift-storage-0\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.195387 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-ovsdbserver-nb\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.196444 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-dns-svc\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.196498 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-ovsdbserver-nb\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.197051 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-config\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.197643 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-dns-swift-storage-0\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.197858 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-ovsdbserver-sb\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.198184 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cf951386-0eef-4a7f-9247-b854b9a39d7b-openstack-edpm-ipam\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.230562 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc7mz\" (UniqueName: \"kubernetes.io/projected/cf951386-0eef-4a7f-9247-b854b9a39d7b-kube-api-access-pc7mz\") pod \"dnsmasq-dns-585f78d44c-nlbbz\" (UID: \"cf951386-0eef-4a7f-9247-b854b9a39d7b\") " pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.316776 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.351662 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.411715 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-nb\") pod \"10c33784-e317-418b-bdf5-ea60acd29d98\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.412414 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-sb\") pod \"10c33784-e317-418b-bdf5-ea60acd29d98\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.412486 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-swift-storage-0\") pod \"10c33784-e317-418b-bdf5-ea60acd29d98\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.412523 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-config\") pod \"10c33784-e317-418b-bdf5-ea60acd29d98\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.412551 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drnbg\" (UniqueName: \"kubernetes.io/projected/10c33784-e317-418b-bdf5-ea60acd29d98-kube-api-access-drnbg\") pod \"10c33784-e317-418b-bdf5-ea60acd29d98\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.412582 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-svc\") pod \"10c33784-e317-418b-bdf5-ea60acd29d98\" (UID: \"10c33784-e317-418b-bdf5-ea60acd29d98\") " Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.444596 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10c33784-e317-418b-bdf5-ea60acd29d98-kube-api-access-drnbg" (OuterVolumeSpecName: "kube-api-access-drnbg") pod "10c33784-e317-418b-bdf5-ea60acd29d98" (UID: "10c33784-e317-418b-bdf5-ea60acd29d98"). InnerVolumeSpecName "kube-api-access-drnbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.508970 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "10c33784-e317-418b-bdf5-ea60acd29d98" (UID: "10c33784-e317-418b-bdf5-ea60acd29d98"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.516523 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.516560 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drnbg\" (UniqueName: \"kubernetes.io/projected/10c33784-e317-418b-bdf5-ea60acd29d98-kube-api-access-drnbg\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.533041 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "10c33784-e317-418b-bdf5-ea60acd29d98" (UID: "10c33784-e317-418b-bdf5-ea60acd29d98"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.560707 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "10c33784-e317-418b-bdf5-ea60acd29d98" (UID: "10c33784-e317-418b-bdf5-ea60acd29d98"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.608058 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-config" (OuterVolumeSpecName: "config") pod "10c33784-e317-418b-bdf5-ea60acd29d98" (UID: "10c33784-e317-418b-bdf5-ea60acd29d98"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.610509 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "10c33784-e317-418b-bdf5-ea60acd29d98" (UID: "10c33784-e317-418b-bdf5-ea60acd29d98"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.619794 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.619846 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.619856 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.619869 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10c33784-e317-418b-bdf5-ea60acd29d98-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.731577 4810 generic.go:334] "Generic (PLEG): container finished" podID="10c33784-e317-418b-bdf5-ea60acd29d98" containerID="24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee" exitCode=0 Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.731633 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" event={"ID":"10c33784-e317-418b-bdf5-ea60acd29d98","Type":"ContainerDied","Data":"24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee"} Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.731649 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.731688 4810 scope.go:117] "RemoveContainer" containerID="24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.731661 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" event={"ID":"10c33784-e317-418b-bdf5-ea60acd29d98","Type":"ContainerDied","Data":"4417d457ba459e14102f801da08e5dc17d73201dd10f265a6ffb902212d4edbd"} Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.789044 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb9f56f5-cg226"] Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.805636 4810 scope.go:117] "RemoveContainer" containerID="0d3d5c82441542254298d6151a1d8291d9d91d20d33a937c53f8c8f60ad88742" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.806641 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb9f56f5-cg226"] Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.850872 4810 scope.go:117] "RemoveContainer" containerID="24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee" Sep 30 08:26:20 crc kubenswrapper[4810]: E0930 08:26:20.851457 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee\": container with ID starting with 24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee not found: ID does not exist" containerID="24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.851490 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee"} err="failed to get container status \"24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee\": rpc error: code = NotFound desc = could not find container \"24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee\": container with ID starting with 24eb6c5db3d845a736bac2452f98ee581f594534f1dec96c1bf3b50c9c9e0aee not found: ID does not exist" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.851512 4810 scope.go:117] "RemoveContainer" containerID="0d3d5c82441542254298d6151a1d8291d9d91d20d33a937c53f8c8f60ad88742" Sep 30 08:26:20 crc kubenswrapper[4810]: E0930 08:26:20.851907 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d3d5c82441542254298d6151a1d8291d9d91d20d33a937c53f8c8f60ad88742\": container with ID starting with 0d3d5c82441542254298d6151a1d8291d9d91d20d33a937c53f8c8f60ad88742 not found: ID does not exist" containerID="0d3d5c82441542254298d6151a1d8291d9d91d20d33a937c53f8c8f60ad88742" Sep 30 08:26:20 crc kubenswrapper[4810]: I0930 08:26:20.851942 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d3d5c82441542254298d6151a1d8291d9d91d20d33a937c53f8c8f60ad88742"} err="failed to get container status \"0d3d5c82441542254298d6151a1d8291d9d91d20d33a937c53f8c8f60ad88742\": rpc error: code = NotFound desc = could not find container \"0d3d5c82441542254298d6151a1d8291d9d91d20d33a937c53f8c8f60ad88742\": container with ID starting with 0d3d5c82441542254298d6151a1d8291d9d91d20d33a937c53f8c8f60ad88742 not found: ID does not exist" Sep 30 08:26:21 crc kubenswrapper[4810]: W0930 08:26:21.024972 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf951386_0eef_4a7f_9247_b854b9a39d7b.slice/crio-3bc9e9fca6c6bc7d82c5bd1a455948b1d4a2ef15027aa148e1d11b9c1dea2826 WatchSource:0}: Error finding container 3bc9e9fca6c6bc7d82c5bd1a455948b1d4a2ef15027aa148e1d11b9c1dea2826: Status 404 returned error can't find the container with id 3bc9e9fca6c6bc7d82c5bd1a455948b1d4a2ef15027aa148e1d11b9c1dea2826 Sep 30 08:26:21 crc kubenswrapper[4810]: I0930 08:26:21.030258 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-585f78d44c-nlbbz"] Sep 30 08:26:21 crc kubenswrapper[4810]: I0930 08:26:21.317704 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10c33784-e317-418b-bdf5-ea60acd29d98" path="/var/lib/kubelet/pods/10c33784-e317-418b-bdf5-ea60acd29d98/volumes" Sep 30 08:26:21 crc kubenswrapper[4810]: I0930 08:26:21.745392 4810 generic.go:334] "Generic (PLEG): container finished" podID="cf951386-0eef-4a7f-9247-b854b9a39d7b" containerID="aa9b730a6c9c31f133df9a2357c8123a1907962aae716cf31e89b5c33ebf512e" exitCode=0 Sep 30 08:26:21 crc kubenswrapper[4810]: I0930 08:26:21.745435 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" event={"ID":"cf951386-0eef-4a7f-9247-b854b9a39d7b","Type":"ContainerDied","Data":"aa9b730a6c9c31f133df9a2357c8123a1907962aae716cf31e89b5c33ebf512e"} Sep 30 08:26:21 crc kubenswrapper[4810]: I0930 08:26:21.745458 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" event={"ID":"cf951386-0eef-4a7f-9247-b854b9a39d7b","Type":"ContainerStarted","Data":"3bc9e9fca6c6bc7d82c5bd1a455948b1d4a2ef15027aa148e1d11b9c1dea2826"} Sep 30 08:26:22 crc kubenswrapper[4810]: I0930 08:26:22.762988 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" event={"ID":"cf951386-0eef-4a7f-9247-b854b9a39d7b","Type":"ContainerStarted","Data":"d7ec7954eceeb11fc6a902f2a8e1c989dd1b6d36b6f4a1c3d4fd7b207e5882b9"} Sep 30 08:26:22 crc kubenswrapper[4810]: I0930 08:26:22.763621 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:22 crc kubenswrapper[4810]: I0930 08:26:22.804815 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" podStartSLOduration=3.8047892880000003 podStartE2EDuration="3.804789288s" podCreationTimestamp="2025-09-30 08:26:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:26:22.794578022 +0000 UTC m=+1406.246777309" watchObservedRunningTime="2025-09-30 08:26:22.804789288 +0000 UTC m=+1406.256988565" Sep 30 08:26:23 crc kubenswrapper[4810]: I0930 08:26:23.783647 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:23 crc kubenswrapper[4810]: I0930 08:26:23.785366 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:23 crc kubenswrapper[4810]: I0930 08:26:23.852422 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:24 crc kubenswrapper[4810]: I0930 08:26:24.836176 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:24 crc kubenswrapper[4810]: I0930 08:26:24.895609 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2snhw"] Sep 30 08:26:25 crc kubenswrapper[4810]: I0930 08:26:25.130165 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6bb9f56f5-cg226" podUID="10c33784-e317-418b-bdf5-ea60acd29d98" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.222:5353: i/o timeout" Sep 30 08:26:26 crc kubenswrapper[4810]: I0930 08:26:26.821621 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2snhw" podUID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" containerName="registry-server" containerID="cri-o://68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c" gracePeriod=2 Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.402692 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.507905 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-75xrr"] Sep 30 08:26:27 crc kubenswrapper[4810]: E0930 08:26:27.508681 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c33784-e317-418b-bdf5-ea60acd29d98" containerName="init" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.508695 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c33784-e317-418b-bdf5-ea60acd29d98" containerName="init" Sep 30 08:26:27 crc kubenswrapper[4810]: E0930 08:26:27.508714 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c33784-e317-418b-bdf5-ea60acd29d98" containerName="dnsmasq-dns" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.508721 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c33784-e317-418b-bdf5-ea60acd29d98" containerName="dnsmasq-dns" Sep 30 08:26:27 crc kubenswrapper[4810]: E0930 08:26:27.508734 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" containerName="registry-server" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.508740 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" containerName="registry-server" Sep 30 08:26:27 crc kubenswrapper[4810]: E0930 08:26:27.508767 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" containerName="extract-content" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.508773 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" containerName="extract-content" Sep 30 08:26:27 crc kubenswrapper[4810]: E0930 08:26:27.508783 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" containerName="extract-utilities" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.508788 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" containerName="extract-utilities" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.508973 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" containerName="registry-server" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.508987 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c33784-e317-418b-bdf5-ea60acd29d98" containerName="dnsmasq-dns" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.510487 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.526049 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-75xrr"] Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.580234 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-catalog-content\") pod \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.580472 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4mpp\" (UniqueName: \"kubernetes.io/projected/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-kube-api-access-c4mpp\") pod \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.580514 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-utilities\") pod \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\" (UID: \"e10b0680-a0ad-45dd-8abf-10f91f4aba4c\") " Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.581445 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-utilities" (OuterVolumeSpecName: "utilities") pod "e10b0680-a0ad-45dd-8abf-10f91f4aba4c" (UID: "e10b0680-a0ad-45dd-8abf-10f91f4aba4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.587946 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-kube-api-access-c4mpp" (OuterVolumeSpecName: "kube-api-access-c4mpp") pod "e10b0680-a0ad-45dd-8abf-10f91f4aba4c" (UID: "e10b0680-a0ad-45dd-8abf-10f91f4aba4c"). InnerVolumeSpecName "kube-api-access-c4mpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.626960 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e10b0680-a0ad-45dd-8abf-10f91f4aba4c" (UID: "e10b0680-a0ad-45dd-8abf-10f91f4aba4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.683653 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-utilities\") pod \"redhat-operators-75xrr\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.683888 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-catalog-content\") pod \"redhat-operators-75xrr\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.683958 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5pbx\" (UniqueName: \"kubernetes.io/projected/8fbf7c16-8de8-46c7-aef7-222e9d70935c-kube-api-access-p5pbx\") pod \"redhat-operators-75xrr\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.684032 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4mpp\" (UniqueName: \"kubernetes.io/projected/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-kube-api-access-c4mpp\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.684047 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.684060 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10b0680-a0ad-45dd-8abf-10f91f4aba4c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.786147 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5pbx\" (UniqueName: \"kubernetes.io/projected/8fbf7c16-8de8-46c7-aef7-222e9d70935c-kube-api-access-p5pbx\") pod \"redhat-operators-75xrr\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.786259 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-utilities\") pod \"redhat-operators-75xrr\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.786640 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-catalog-content\") pod \"redhat-operators-75xrr\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.786733 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-utilities\") pod \"redhat-operators-75xrr\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.787102 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-catalog-content\") pod \"redhat-operators-75xrr\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.808089 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5pbx\" (UniqueName: \"kubernetes.io/projected/8fbf7c16-8de8-46c7-aef7-222e9d70935c-kube-api-access-p5pbx\") pod \"redhat-operators-75xrr\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.829930 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.836831 4810 generic.go:334] "Generic (PLEG): container finished" podID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" containerID="68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c" exitCode=0 Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.836901 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2snhw" event={"ID":"e10b0680-a0ad-45dd-8abf-10f91f4aba4c","Type":"ContainerDied","Data":"68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c"} Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.836943 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2snhw" event={"ID":"e10b0680-a0ad-45dd-8abf-10f91f4aba4c","Type":"ContainerDied","Data":"0e6e65e5c17e22416d3a4a85ac7e27fe1eb18e23c72440dc5fc72c00dabcb4e5"} Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.836985 4810 scope.go:117] "RemoveContainer" containerID="68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.837045 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2snhw" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.905225 4810 scope.go:117] "RemoveContainer" containerID="8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f" Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.915507 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2snhw"] Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.923559 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2snhw"] Sep 30 08:26:27 crc kubenswrapper[4810]: I0930 08:26:27.958477 4810 scope.go:117] "RemoveContainer" containerID="6f0a175dabbc3d169b283d81207006f57217968cf6a3cdcd6de59518d2b722c7" Sep 30 08:26:28 crc kubenswrapper[4810]: I0930 08:26:28.032241 4810 scope.go:117] "RemoveContainer" containerID="68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c" Sep 30 08:26:28 crc kubenswrapper[4810]: E0930 08:26:28.033021 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c\": container with ID starting with 68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c not found: ID does not exist" containerID="68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c" Sep 30 08:26:28 crc kubenswrapper[4810]: I0930 08:26:28.033053 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c"} err="failed to get container status \"68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c\": rpc error: code = NotFound desc = could not find container \"68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c\": container with ID starting with 68c0c69e9fbd8d49a6b05a2586983359939c9dc7d7617a289326728a620ac95c not found: ID does not exist" Sep 30 08:26:28 crc kubenswrapper[4810]: I0930 08:26:28.033075 4810 scope.go:117] "RemoveContainer" containerID="8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f" Sep 30 08:26:28 crc kubenswrapper[4810]: E0930 08:26:28.033796 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f\": container with ID starting with 8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f not found: ID does not exist" containerID="8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f" Sep 30 08:26:28 crc kubenswrapper[4810]: I0930 08:26:28.033854 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f"} err="failed to get container status \"8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f\": rpc error: code = NotFound desc = could not find container \"8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f\": container with ID starting with 8685a005989aa47a3adb257c7a83ab7bc380166c7fe379708f35f79eba52ed0f not found: ID does not exist" Sep 30 08:26:28 crc kubenswrapper[4810]: I0930 08:26:28.033892 4810 scope.go:117] "RemoveContainer" containerID="6f0a175dabbc3d169b283d81207006f57217968cf6a3cdcd6de59518d2b722c7" Sep 30 08:26:28 crc kubenswrapper[4810]: E0930 08:26:28.034466 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f0a175dabbc3d169b283d81207006f57217968cf6a3cdcd6de59518d2b722c7\": container with ID starting with 6f0a175dabbc3d169b283d81207006f57217968cf6a3cdcd6de59518d2b722c7 not found: ID does not exist" containerID="6f0a175dabbc3d169b283d81207006f57217968cf6a3cdcd6de59518d2b722c7" Sep 30 08:26:28 crc kubenswrapper[4810]: I0930 08:26:28.034571 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f0a175dabbc3d169b283d81207006f57217968cf6a3cdcd6de59518d2b722c7"} err="failed to get container status \"6f0a175dabbc3d169b283d81207006f57217968cf6a3cdcd6de59518d2b722c7\": rpc error: code = NotFound desc = could not find container \"6f0a175dabbc3d169b283d81207006f57217968cf6a3cdcd6de59518d2b722c7\": container with ID starting with 6f0a175dabbc3d169b283d81207006f57217968cf6a3cdcd6de59518d2b722c7 not found: ID does not exist" Sep 30 08:26:28 crc kubenswrapper[4810]: I0930 08:26:28.300365 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-75xrr"] Sep 30 08:26:28 crc kubenswrapper[4810]: I0930 08:26:28.850868 4810 generic.go:334] "Generic (PLEG): container finished" podID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" containerID="3d07e1c0d43e0821e47d489df70bb102cdcf252d62514f71c4c90d3b3f026522" exitCode=0 Sep 30 08:26:28 crc kubenswrapper[4810]: I0930 08:26:28.851010 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75xrr" event={"ID":"8fbf7c16-8de8-46c7-aef7-222e9d70935c","Type":"ContainerDied","Data":"3d07e1c0d43e0821e47d489df70bb102cdcf252d62514f71c4c90d3b3f026522"} Sep 30 08:26:28 crc kubenswrapper[4810]: I0930 08:26:28.852527 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75xrr" event={"ID":"8fbf7c16-8de8-46c7-aef7-222e9d70935c","Type":"ContainerStarted","Data":"52fd3ec5b29ead4a70e2d9973e0534ec99a93323d85b59a87ded0b5cdfd9034d"} Sep 30 08:26:29 crc kubenswrapper[4810]: I0930 08:26:29.322451 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e10b0680-a0ad-45dd-8abf-10f91f4aba4c" path="/var/lib/kubelet/pods/e10b0680-a0ad-45dd-8abf-10f91f4aba4c/volumes" Sep 30 08:26:29 crc kubenswrapper[4810]: I0930 08:26:29.889233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75xrr" event={"ID":"8fbf7c16-8de8-46c7-aef7-222e9d70935c","Type":"ContainerStarted","Data":"71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a"} Sep 30 08:26:30 crc kubenswrapper[4810]: I0930 08:26:30.353512 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-585f78d44c-nlbbz" Sep 30 08:26:30 crc kubenswrapper[4810]: I0930 08:26:30.421892 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cbfdf555-8fpkr"] Sep 30 08:26:30 crc kubenswrapper[4810]: I0930 08:26:30.422351 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" podUID="6a51583a-0d9a-4f85-8768-8ff462be2485" containerName="dnsmasq-dns" containerID="cri-o://1fccb98ff7343b898c2ff0e27bbf9f156471d551d329b7938b1a77fd54308054" gracePeriod=10 Sep 30 08:26:30 crc kubenswrapper[4810]: I0930 08:26:30.907495 4810 generic.go:334] "Generic (PLEG): container finished" podID="6a51583a-0d9a-4f85-8768-8ff462be2485" containerID="1fccb98ff7343b898c2ff0e27bbf9f156471d551d329b7938b1a77fd54308054" exitCode=0 Sep 30 08:26:30 crc kubenswrapper[4810]: I0930 08:26:30.907598 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" event={"ID":"6a51583a-0d9a-4f85-8768-8ff462be2485","Type":"ContainerDied","Data":"1fccb98ff7343b898c2ff0e27bbf9f156471d551d329b7938b1a77fd54308054"} Sep 30 08:26:30 crc kubenswrapper[4810]: I0930 08:26:30.907867 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" event={"ID":"6a51583a-0d9a-4f85-8768-8ff462be2485","Type":"ContainerDied","Data":"73de7f84baf4f9d191c53d9612150ededfae67df86fa51a678168f7147543c40"} Sep 30 08:26:30 crc kubenswrapper[4810]: I0930 08:26:30.907888 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73de7f84baf4f9d191c53d9612150ededfae67df86fa51a678168f7147543c40" Sep 30 08:26:30 crc kubenswrapper[4810]: I0930 08:26:30.912043 4810 generic.go:334] "Generic (PLEG): container finished" podID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" containerID="71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a" exitCode=0 Sep 30 08:26:30 crc kubenswrapper[4810]: I0930 08:26:30.912114 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75xrr" event={"ID":"8fbf7c16-8de8-46c7-aef7-222e9d70935c","Type":"ContainerDied","Data":"71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a"} Sep 30 08:26:30 crc kubenswrapper[4810]: I0930 08:26:30.966398 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.159948 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-svc\") pod \"6a51583a-0d9a-4f85-8768-8ff462be2485\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.160023 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-sb\") pod \"6a51583a-0d9a-4f85-8768-8ff462be2485\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.160045 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-openstack-edpm-ipam\") pod \"6a51583a-0d9a-4f85-8768-8ff462be2485\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.160063 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-config\") pod \"6a51583a-0d9a-4f85-8768-8ff462be2485\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.160097 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-nb\") pod \"6a51583a-0d9a-4f85-8768-8ff462be2485\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.160202 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-swift-storage-0\") pod \"6a51583a-0d9a-4f85-8768-8ff462be2485\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.160247 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pxh4\" (UniqueName: \"kubernetes.io/projected/6a51583a-0d9a-4f85-8768-8ff462be2485-kube-api-access-9pxh4\") pod \"6a51583a-0d9a-4f85-8768-8ff462be2485\" (UID: \"6a51583a-0d9a-4f85-8768-8ff462be2485\") " Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.171128 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a51583a-0d9a-4f85-8768-8ff462be2485-kube-api-access-9pxh4" (OuterVolumeSpecName: "kube-api-access-9pxh4") pod "6a51583a-0d9a-4f85-8768-8ff462be2485" (UID: "6a51583a-0d9a-4f85-8768-8ff462be2485"). InnerVolumeSpecName "kube-api-access-9pxh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.219544 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6a51583a-0d9a-4f85-8768-8ff462be2485" (UID: "6a51583a-0d9a-4f85-8768-8ff462be2485"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.224145 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "6a51583a-0d9a-4f85-8768-8ff462be2485" (UID: "6a51583a-0d9a-4f85-8768-8ff462be2485"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.230091 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-config" (OuterVolumeSpecName: "config") pod "6a51583a-0d9a-4f85-8768-8ff462be2485" (UID: "6a51583a-0d9a-4f85-8768-8ff462be2485"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.233809 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6a51583a-0d9a-4f85-8768-8ff462be2485" (UID: "6a51583a-0d9a-4f85-8768-8ff462be2485"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.260009 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6a51583a-0d9a-4f85-8768-8ff462be2485" (UID: "6a51583a-0d9a-4f85-8768-8ff462be2485"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.265025 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.265058 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pxh4\" (UniqueName: \"kubernetes.io/projected/6a51583a-0d9a-4f85-8768-8ff462be2485-kube-api-access-9pxh4\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.265071 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.265082 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.265094 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.265107 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.269731 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6a51583a-0d9a-4f85-8768-8ff462be2485" (UID: "6a51583a-0d9a-4f85-8768-8ff462be2485"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.366928 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a51583a-0d9a-4f85-8768-8ff462be2485-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.926399 4810 generic.go:334] "Generic (PLEG): container finished" podID="b695439c-7bde-408e-b90a-0aa0ff0bc494" containerID="94380eeeebea55ed64cd5a48169987341384a32a723ebc853581689bbfa19b46" exitCode=0 Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.926523 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b695439c-7bde-408e-b90a-0aa0ff0bc494","Type":"ContainerDied","Data":"94380eeeebea55ed64cd5a48169987341384a32a723ebc853581689bbfa19b46"} Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.930308 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75xrr" event={"ID":"8fbf7c16-8de8-46c7-aef7-222e9d70935c","Type":"ContainerStarted","Data":"6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b"} Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.931986 4810 generic.go:334] "Generic (PLEG): container finished" podID="d0344203-4b62-409c-af24-0e619dfad39a" containerID="bc619bbf4dff258f9e44808baef181ad87cdafbb25c295a7d6dd54b8bee3f2a3" exitCode=0 Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.932079 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cbfdf555-8fpkr" Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.932080 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d0344203-4b62-409c-af24-0e619dfad39a","Type":"ContainerDied","Data":"bc619bbf4dff258f9e44808baef181ad87cdafbb25c295a7d6dd54b8bee3f2a3"} Sep 30 08:26:31 crc kubenswrapper[4810]: I0930 08:26:31.993806 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-75xrr" podStartSLOduration=2.2093727579999998 podStartE2EDuration="4.993779109s" podCreationTimestamp="2025-09-30 08:26:27 +0000 UTC" firstStartedPulling="2025-09-30 08:26:28.853609042 +0000 UTC m=+1412.305808309" lastFinishedPulling="2025-09-30 08:26:31.638015393 +0000 UTC m=+1415.090214660" observedRunningTime="2025-09-30 08:26:31.980764113 +0000 UTC m=+1415.432963370" watchObservedRunningTime="2025-09-30 08:26:31.993779109 +0000 UTC m=+1415.445978386" Sep 30 08:26:32 crc kubenswrapper[4810]: I0930 08:26:32.149820 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cbfdf555-8fpkr"] Sep 30 08:26:32 crc kubenswrapper[4810]: I0930 08:26:32.159629 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cbfdf555-8fpkr"] Sep 30 08:26:32 crc kubenswrapper[4810]: I0930 08:26:32.942759 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b695439c-7bde-408e-b90a-0aa0ff0bc494","Type":"ContainerStarted","Data":"758a8f4ee7816bbdf4a5f3f44b2fbcf4a85773587430ae2604b42ee93ad1b0c8"} Sep 30 08:26:32 crc kubenswrapper[4810]: I0930 08:26:32.943204 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 30 08:26:32 crc kubenswrapper[4810]: I0930 08:26:32.945307 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d0344203-4b62-409c-af24-0e619dfad39a","Type":"ContainerStarted","Data":"17b6b41af1ccdf6067c6be1bd4e475b8acf215cf61dde9c016d69076dfc477c5"} Sep 30 08:26:32 crc kubenswrapper[4810]: I0930 08:26:32.986371 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=33.986349974 podStartE2EDuration="33.986349974s" podCreationTimestamp="2025-09-30 08:25:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:26:32.978041403 +0000 UTC m=+1416.430240670" watchObservedRunningTime="2025-09-30 08:26:32.986349974 +0000 UTC m=+1416.438549241" Sep 30 08:26:33 crc kubenswrapper[4810]: I0930 08:26:33.013327 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=34.013310184 podStartE2EDuration="34.013310184s" podCreationTimestamp="2025-09-30 08:25:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:26:33.011692727 +0000 UTC m=+1416.463892004" watchObservedRunningTime="2025-09-30 08:26:33.013310184 +0000 UTC m=+1416.465509451" Sep 30 08:26:33 crc kubenswrapper[4810]: I0930 08:26:33.320997 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a51583a-0d9a-4f85-8768-8ff462be2485" path="/var/lib/kubelet/pods/6a51583a-0d9a-4f85-8768-8ff462be2485/volumes" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.709015 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjmj"] Sep 30 08:26:34 crc kubenswrapper[4810]: E0930 08:26:34.709568 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a51583a-0d9a-4f85-8768-8ff462be2485" containerName="dnsmasq-dns" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.709587 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a51583a-0d9a-4f85-8768-8ff462be2485" containerName="dnsmasq-dns" Sep 30 08:26:34 crc kubenswrapper[4810]: E0930 08:26:34.709603 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a51583a-0d9a-4f85-8768-8ff462be2485" containerName="init" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.709611 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a51583a-0d9a-4f85-8768-8ff462be2485" containerName="init" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.709848 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a51583a-0d9a-4f85-8768-8ff462be2485" containerName="dnsmasq-dns" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.712547 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.726405 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjmj"] Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.832174 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-utilities\") pod \"redhat-marketplace-qcjmj\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.832375 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc6j2\" (UniqueName: \"kubernetes.io/projected/aea6b294-b4a3-4e84-a63c-cd3a18a74229-kube-api-access-bc6j2\") pod \"redhat-marketplace-qcjmj\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.832444 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-catalog-content\") pod \"redhat-marketplace-qcjmj\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.935007 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc6j2\" (UniqueName: \"kubernetes.io/projected/aea6b294-b4a3-4e84-a63c-cd3a18a74229-kube-api-access-bc6j2\") pod \"redhat-marketplace-qcjmj\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.935483 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-catalog-content\") pod \"redhat-marketplace-qcjmj\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.935676 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-utilities\") pod \"redhat-marketplace-qcjmj\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.935929 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-catalog-content\") pod \"redhat-marketplace-qcjmj\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.936149 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-utilities\") pod \"redhat-marketplace-qcjmj\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:34 crc kubenswrapper[4810]: I0930 08:26:34.968704 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc6j2\" (UniqueName: \"kubernetes.io/projected/aea6b294-b4a3-4e84-a63c-cd3a18a74229-kube-api-access-bc6j2\") pod \"redhat-marketplace-qcjmj\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:35 crc kubenswrapper[4810]: I0930 08:26:35.041712 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:35 crc kubenswrapper[4810]: W0930 08:26:35.541322 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaea6b294_b4a3_4e84_a63c_cd3a18a74229.slice/crio-2a3afac375702ca547beaf45a97ac0745d8628ae977c88ebdc69408214872e40 WatchSource:0}: Error finding container 2a3afac375702ca547beaf45a97ac0745d8628ae977c88ebdc69408214872e40: Status 404 returned error can't find the container with id 2a3afac375702ca547beaf45a97ac0745d8628ae977c88ebdc69408214872e40 Sep 30 08:26:35 crc kubenswrapper[4810]: I0930 08:26:35.545109 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjmj"] Sep 30 08:26:35 crc kubenswrapper[4810]: I0930 08:26:35.978319 4810 generic.go:334] "Generic (PLEG): container finished" podID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" containerID="f86f3b127690ce87a477ceca9b0a7c858b464eb94f4fe5daa67797e99048cae3" exitCode=0 Sep 30 08:26:35 crc kubenswrapper[4810]: I0930 08:26:35.978386 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjmj" event={"ID":"aea6b294-b4a3-4e84-a63c-cd3a18a74229","Type":"ContainerDied","Data":"f86f3b127690ce87a477ceca9b0a7c858b464eb94f4fe5daa67797e99048cae3"} Sep 30 08:26:35 crc kubenswrapper[4810]: I0930 08:26:35.978416 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjmj" event={"ID":"aea6b294-b4a3-4e84-a63c-cd3a18a74229","Type":"ContainerStarted","Data":"2a3afac375702ca547beaf45a97ac0745d8628ae977c88ebdc69408214872e40"} Sep 30 08:26:37 crc kubenswrapper[4810]: I0930 08:26:37.830825 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:37 crc kubenswrapper[4810]: I0930 08:26:37.831165 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:37 crc kubenswrapper[4810]: I0930 08:26:37.904953 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:38 crc kubenswrapper[4810]: I0930 08:26:38.015613 4810 generic.go:334] "Generic (PLEG): container finished" podID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" containerID="3d3f2e24ed21f1cad54352b32b02b9313f9d0c9e8992b858246318510992ceef" exitCode=0 Sep 30 08:26:38 crc kubenswrapper[4810]: I0930 08:26:38.015703 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjmj" event={"ID":"aea6b294-b4a3-4e84-a63c-cd3a18a74229","Type":"ContainerDied","Data":"3d3f2e24ed21f1cad54352b32b02b9313f9d0c9e8992b858246318510992ceef"} Sep 30 08:26:38 crc kubenswrapper[4810]: I0930 08:26:38.085730 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:39 crc kubenswrapper[4810]: I0930 08:26:39.029233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjmj" event={"ID":"aea6b294-b4a3-4e84-a63c-cd3a18a74229","Type":"ContainerStarted","Data":"28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502"} Sep 30 08:26:39 crc kubenswrapper[4810]: I0930 08:26:39.051037 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qcjmj" podStartSLOduration=2.503320746 podStartE2EDuration="5.051014367s" podCreationTimestamp="2025-09-30 08:26:34 +0000 UTC" firstStartedPulling="2025-09-30 08:26:35.97995455 +0000 UTC m=+1419.432153817" lastFinishedPulling="2025-09-30 08:26:38.527648171 +0000 UTC m=+1421.979847438" observedRunningTime="2025-09-30 08:26:39.047724742 +0000 UTC m=+1422.499924009" watchObservedRunningTime="2025-09-30 08:26:39.051014367 +0000 UTC m=+1422.503213654" Sep 30 08:26:39 crc kubenswrapper[4810]: I0930 08:26:39.702107 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-75xrr"] Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.038475 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-75xrr" podUID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" containerName="registry-server" containerID="cri-o://6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b" gracePeriod=2 Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.185385 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.574526 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.658698 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5pbx\" (UniqueName: \"kubernetes.io/projected/8fbf7c16-8de8-46c7-aef7-222e9d70935c-kube-api-access-p5pbx\") pod \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.658791 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-catalog-content\") pod \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.658913 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-utilities\") pod \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\" (UID: \"8fbf7c16-8de8-46c7-aef7-222e9d70935c\") " Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.659800 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-utilities" (OuterVolumeSpecName: "utilities") pod "8fbf7c16-8de8-46c7-aef7-222e9d70935c" (UID: "8fbf7c16-8de8-46c7-aef7-222e9d70935c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.685979 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fbf7c16-8de8-46c7-aef7-222e9d70935c-kube-api-access-p5pbx" (OuterVolumeSpecName: "kube-api-access-p5pbx") pod "8fbf7c16-8de8-46c7-aef7-222e9d70935c" (UID: "8fbf7c16-8de8-46c7-aef7-222e9d70935c"). InnerVolumeSpecName "kube-api-access-p5pbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.761304 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5pbx\" (UniqueName: \"kubernetes.io/projected/8fbf7c16-8de8-46c7-aef7-222e9d70935c-kube-api-access-p5pbx\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.761338 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.775128 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fbf7c16-8de8-46c7-aef7-222e9d70935c" (UID: "8fbf7c16-8de8-46c7-aef7-222e9d70935c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:26:40 crc kubenswrapper[4810]: I0930 08:26:40.863584 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7c16-8de8-46c7-aef7-222e9d70935c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.050521 4810 generic.go:334] "Generic (PLEG): container finished" podID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" containerID="6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b" exitCode=0 Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.050596 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75xrr" event={"ID":"8fbf7c16-8de8-46c7-aef7-222e9d70935c","Type":"ContainerDied","Data":"6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b"} Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.050632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75xrr" event={"ID":"8fbf7c16-8de8-46c7-aef7-222e9d70935c","Type":"ContainerDied","Data":"52fd3ec5b29ead4a70e2d9973e0534ec99a93323d85b59a87ded0b5cdfd9034d"} Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.050652 4810 scope.go:117] "RemoveContainer" containerID="6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b" Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.050679 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75xrr" Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.084816 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-75xrr"] Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.094520 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-75xrr"] Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.097595 4810 scope.go:117] "RemoveContainer" containerID="71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a" Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.155578 4810 scope.go:117] "RemoveContainer" containerID="3d07e1c0d43e0821e47d489df70bb102cdcf252d62514f71c4c90d3b3f026522" Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.198395 4810 scope.go:117] "RemoveContainer" containerID="6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b" Sep 30 08:26:41 crc kubenswrapper[4810]: E0930 08:26:41.199492 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b\": container with ID starting with 6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b not found: ID does not exist" containerID="6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b" Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.199554 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b"} err="failed to get container status \"6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b\": rpc error: code = NotFound desc = could not find container \"6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b\": container with ID starting with 6643c109e8628276dd2012ee36e7062f0c257f30103ac010e30fe42426b66d7b not found: ID does not exist" Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.199583 4810 scope.go:117] "RemoveContainer" containerID="71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a" Sep 30 08:26:41 crc kubenswrapper[4810]: E0930 08:26:41.200079 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a\": container with ID starting with 71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a not found: ID does not exist" containerID="71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a" Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.200113 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a"} err="failed to get container status \"71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a\": rpc error: code = NotFound desc = could not find container \"71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a\": container with ID starting with 71501dcd1dcfb5b6c70ce88dd753e286b749621ecc1e678fde6ac48dad03f30a not found: ID does not exist" Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.200144 4810 scope.go:117] "RemoveContainer" containerID="3d07e1c0d43e0821e47d489df70bb102cdcf252d62514f71c4c90d3b3f026522" Sep 30 08:26:41 crc kubenswrapper[4810]: E0930 08:26:41.201582 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d07e1c0d43e0821e47d489df70bb102cdcf252d62514f71c4c90d3b3f026522\": container with ID starting with 3d07e1c0d43e0821e47d489df70bb102cdcf252d62514f71c4c90d3b3f026522 not found: ID does not exist" containerID="3d07e1c0d43e0821e47d489df70bb102cdcf252d62514f71c4c90d3b3f026522" Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.201609 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d07e1c0d43e0821e47d489df70bb102cdcf252d62514f71c4c90d3b3f026522"} err="failed to get container status \"3d07e1c0d43e0821e47d489df70bb102cdcf252d62514f71c4c90d3b3f026522\": rpc error: code = NotFound desc = could not find container \"3d07e1c0d43e0821e47d489df70bb102cdcf252d62514f71c4c90d3b3f026522\": container with ID starting with 3d07e1c0d43e0821e47d489df70bb102cdcf252d62514f71c4c90d3b3f026522 not found: ID does not exist" Sep 30 08:26:41 crc kubenswrapper[4810]: I0930 08:26:41.317482 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" path="/var/lib/kubelet/pods/8fbf7c16-8de8-46c7-aef7-222e9d70935c/volumes" Sep 30 08:26:44 crc kubenswrapper[4810]: I0930 08:26:44.894259 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r"] Sep 30 08:26:44 crc kubenswrapper[4810]: E0930 08:26:44.895377 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" containerName="extract-content" Sep 30 08:26:44 crc kubenswrapper[4810]: I0930 08:26:44.895395 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" containerName="extract-content" Sep 30 08:26:44 crc kubenswrapper[4810]: E0930 08:26:44.895408 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" containerName="registry-server" Sep 30 08:26:44 crc kubenswrapper[4810]: I0930 08:26:44.895415 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" containerName="registry-server" Sep 30 08:26:44 crc kubenswrapper[4810]: E0930 08:26:44.895444 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" containerName="extract-utilities" Sep 30 08:26:44 crc kubenswrapper[4810]: I0930 08:26:44.895452 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" containerName="extract-utilities" Sep 30 08:26:44 crc kubenswrapper[4810]: I0930 08:26:44.895682 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fbf7c16-8de8-46c7-aef7-222e9d70935c" containerName="registry-server" Sep 30 08:26:44 crc kubenswrapper[4810]: I0930 08:26:44.896546 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:44 crc kubenswrapper[4810]: I0930 08:26:44.900222 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:26:44 crc kubenswrapper[4810]: I0930 08:26:44.900587 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:26:44 crc kubenswrapper[4810]: I0930 08:26:44.900608 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:26:44 crc kubenswrapper[4810]: I0930 08:26:44.900611 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:26:44 crc kubenswrapper[4810]: I0930 08:26:44.907097 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r"] Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.042150 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.043906 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.065460 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.065603 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.065735 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6spw\" (UniqueName: \"kubernetes.io/projected/93d196dc-dc56-4116-bb59-d76601947afd-kube-api-access-x6spw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.065795 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.089917 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.146686 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.167152 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.167261 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6spw\" (UniqueName: \"kubernetes.io/projected/93d196dc-dc56-4116-bb59-d76601947afd-kube-api-access-x6spw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.167336 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.167371 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.173552 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.173966 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.174369 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.185894 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6spw\" (UniqueName: \"kubernetes.io/projected/93d196dc-dc56-4116-bb59-d76601947afd-kube-api-access-x6spw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.230449 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.793358 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r"] Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.911738 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:26:45 crc kubenswrapper[4810]: I0930 08:26:45.911794 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:26:46 crc kubenswrapper[4810]: I0930 08:26:46.099703 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" event={"ID":"93d196dc-dc56-4116-bb59-d76601947afd","Type":"ContainerStarted","Data":"2a9c5cc3369984e1f69580a88545a2e1064501722c35a8c871e142d2048a63c3"} Sep 30 08:26:46 crc kubenswrapper[4810]: I0930 08:26:46.107350 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjmj"] Sep 30 08:26:47 crc kubenswrapper[4810]: I0930 08:26:47.117104 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qcjmj" podUID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" containerName="registry-server" containerID="cri-o://28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502" gracePeriod=2 Sep 30 08:26:47 crc kubenswrapper[4810]: I0930 08:26:47.841386 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:47 crc kubenswrapper[4810]: I0930 08:26:47.923036 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc6j2\" (UniqueName: \"kubernetes.io/projected/aea6b294-b4a3-4e84-a63c-cd3a18a74229-kube-api-access-bc6j2\") pod \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " Sep 30 08:26:47 crc kubenswrapper[4810]: I0930 08:26:47.923208 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-catalog-content\") pod \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " Sep 30 08:26:47 crc kubenswrapper[4810]: I0930 08:26:47.923624 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-utilities\") pod \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\" (UID: \"aea6b294-b4a3-4e84-a63c-cd3a18a74229\") " Sep 30 08:26:47 crc kubenswrapper[4810]: I0930 08:26:47.924930 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-utilities" (OuterVolumeSpecName: "utilities") pod "aea6b294-b4a3-4e84-a63c-cd3a18a74229" (UID: "aea6b294-b4a3-4e84-a63c-cd3a18a74229"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:26:47 crc kubenswrapper[4810]: I0930 08:26:47.931678 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aea6b294-b4a3-4e84-a63c-cd3a18a74229-kube-api-access-bc6j2" (OuterVolumeSpecName: "kube-api-access-bc6j2") pod "aea6b294-b4a3-4e84-a63c-cd3a18a74229" (UID: "aea6b294-b4a3-4e84-a63c-cd3a18a74229"). InnerVolumeSpecName "kube-api-access-bc6j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:26:47 crc kubenswrapper[4810]: I0930 08:26:47.939041 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aea6b294-b4a3-4e84-a63c-cd3a18a74229" (UID: "aea6b294-b4a3-4e84-a63c-cd3a18a74229"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.027335 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.027424 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aea6b294-b4a3-4e84-a63c-cd3a18a74229-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.027435 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc6j2\" (UniqueName: \"kubernetes.io/projected/aea6b294-b4a3-4e84-a63c-cd3a18a74229-kube-api-access-bc6j2\") on node \"crc\" DevicePath \"\"" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.136669 4810 generic.go:334] "Generic (PLEG): container finished" podID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" containerID="28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502" exitCode=0 Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.136713 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjmj" event={"ID":"aea6b294-b4a3-4e84-a63c-cd3a18a74229","Type":"ContainerDied","Data":"28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502"} Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.136742 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjmj" event={"ID":"aea6b294-b4a3-4e84-a63c-cd3a18a74229","Type":"ContainerDied","Data":"2a3afac375702ca547beaf45a97ac0745d8628ae977c88ebdc69408214872e40"} Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.136759 4810 scope.go:117] "RemoveContainer" containerID="28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.136887 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcjmj" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.171410 4810 scope.go:117] "RemoveContainer" containerID="3d3f2e24ed21f1cad54352b32b02b9313f9d0c9e8992b858246318510992ceef" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.185300 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjmj"] Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.198389 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjmj"] Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.217023 4810 scope.go:117] "RemoveContainer" containerID="f86f3b127690ce87a477ceca9b0a7c858b464eb94f4fe5daa67797e99048cae3" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.267837 4810 scope.go:117] "RemoveContainer" containerID="28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502" Sep 30 08:26:48 crc kubenswrapper[4810]: E0930 08:26:48.268484 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502\": container with ID starting with 28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502 not found: ID does not exist" containerID="28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.268536 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502"} err="failed to get container status \"28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502\": rpc error: code = NotFound desc = could not find container \"28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502\": container with ID starting with 28e37ecd514f1f44c201c50a32cd34a5a01f4522572ef531206c99ee3f992502 not found: ID does not exist" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.268566 4810 scope.go:117] "RemoveContainer" containerID="3d3f2e24ed21f1cad54352b32b02b9313f9d0c9e8992b858246318510992ceef" Sep 30 08:26:48 crc kubenswrapper[4810]: E0930 08:26:48.268948 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d3f2e24ed21f1cad54352b32b02b9313f9d0c9e8992b858246318510992ceef\": container with ID starting with 3d3f2e24ed21f1cad54352b32b02b9313f9d0c9e8992b858246318510992ceef not found: ID does not exist" containerID="3d3f2e24ed21f1cad54352b32b02b9313f9d0c9e8992b858246318510992ceef" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.268987 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d3f2e24ed21f1cad54352b32b02b9313f9d0c9e8992b858246318510992ceef"} err="failed to get container status \"3d3f2e24ed21f1cad54352b32b02b9313f9d0c9e8992b858246318510992ceef\": rpc error: code = NotFound desc = could not find container \"3d3f2e24ed21f1cad54352b32b02b9313f9d0c9e8992b858246318510992ceef\": container with ID starting with 3d3f2e24ed21f1cad54352b32b02b9313f9d0c9e8992b858246318510992ceef not found: ID does not exist" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.269014 4810 scope.go:117] "RemoveContainer" containerID="f86f3b127690ce87a477ceca9b0a7c858b464eb94f4fe5daa67797e99048cae3" Sep 30 08:26:48 crc kubenswrapper[4810]: E0930 08:26:48.269618 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f86f3b127690ce87a477ceca9b0a7c858b464eb94f4fe5daa67797e99048cae3\": container with ID starting with f86f3b127690ce87a477ceca9b0a7c858b464eb94f4fe5daa67797e99048cae3 not found: ID does not exist" containerID="f86f3b127690ce87a477ceca9b0a7c858b464eb94f4fe5daa67797e99048cae3" Sep 30 08:26:48 crc kubenswrapper[4810]: I0930 08:26:48.269652 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f86f3b127690ce87a477ceca9b0a7c858b464eb94f4fe5daa67797e99048cae3"} err="failed to get container status \"f86f3b127690ce87a477ceca9b0a7c858b464eb94f4fe5daa67797e99048cae3\": rpc error: code = NotFound desc = could not find container \"f86f3b127690ce87a477ceca9b0a7c858b464eb94f4fe5daa67797e99048cae3\": container with ID starting with f86f3b127690ce87a477ceca9b0a7c858b464eb94f4fe5daa67797e99048cae3 not found: ID does not exist" Sep 30 08:26:49 crc kubenswrapper[4810]: I0930 08:26:49.324522 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" path="/var/lib/kubelet/pods/aea6b294-b4a3-4e84-a63c-cd3a18a74229/volumes" Sep 30 08:26:50 crc kubenswrapper[4810]: I0930 08:26:50.190662 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 30 08:26:50 crc kubenswrapper[4810]: I0930 08:26:50.226456 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 30 08:26:57 crc kubenswrapper[4810]: I0930 08:26:57.256449 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" event={"ID":"93d196dc-dc56-4116-bb59-d76601947afd","Type":"ContainerStarted","Data":"a55094e205c77505d4b8b0630c210519a978719b7d8f8a16baa249b0df7cc4ac"} Sep 30 08:26:57 crc kubenswrapper[4810]: I0930 08:26:57.280577 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" podStartSLOduration=2.187154999 podStartE2EDuration="13.280556015s" podCreationTimestamp="2025-09-30 08:26:44 +0000 UTC" firstStartedPulling="2025-09-30 08:26:45.79085867 +0000 UTC m=+1429.243057937" lastFinishedPulling="2025-09-30 08:26:56.884259656 +0000 UTC m=+1440.336458953" observedRunningTime="2025-09-30 08:26:57.277252469 +0000 UTC m=+1440.729451746" watchObservedRunningTime="2025-09-30 08:26:57.280556015 +0000 UTC m=+1440.732755282" Sep 30 08:27:10 crc kubenswrapper[4810]: I0930 08:27:10.416574 4810 generic.go:334] "Generic (PLEG): container finished" podID="93d196dc-dc56-4116-bb59-d76601947afd" containerID="a55094e205c77505d4b8b0630c210519a978719b7d8f8a16baa249b0df7cc4ac" exitCode=0 Sep 30 08:27:10 crc kubenswrapper[4810]: I0930 08:27:10.416914 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" event={"ID":"93d196dc-dc56-4116-bb59-d76601947afd","Type":"ContainerDied","Data":"a55094e205c77505d4b8b0630c210519a978719b7d8f8a16baa249b0df7cc4ac"} Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.087015 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.256985 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-repo-setup-combined-ca-bundle\") pod \"93d196dc-dc56-4116-bb59-d76601947afd\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.257420 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-inventory\") pod \"93d196dc-dc56-4116-bb59-d76601947afd\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.257655 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-ssh-key\") pod \"93d196dc-dc56-4116-bb59-d76601947afd\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.257889 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6spw\" (UniqueName: \"kubernetes.io/projected/93d196dc-dc56-4116-bb59-d76601947afd-kube-api-access-x6spw\") pod \"93d196dc-dc56-4116-bb59-d76601947afd\" (UID: \"93d196dc-dc56-4116-bb59-d76601947afd\") " Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.264471 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "93d196dc-dc56-4116-bb59-d76601947afd" (UID: "93d196dc-dc56-4116-bb59-d76601947afd"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.276002 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93d196dc-dc56-4116-bb59-d76601947afd-kube-api-access-x6spw" (OuterVolumeSpecName: "kube-api-access-x6spw") pod "93d196dc-dc56-4116-bb59-d76601947afd" (UID: "93d196dc-dc56-4116-bb59-d76601947afd"). InnerVolumeSpecName "kube-api-access-x6spw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.288509 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-inventory" (OuterVolumeSpecName: "inventory") pod "93d196dc-dc56-4116-bb59-d76601947afd" (UID: "93d196dc-dc56-4116-bb59-d76601947afd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.290182 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "93d196dc-dc56-4116-bb59-d76601947afd" (UID: "93d196dc-dc56-4116-bb59-d76601947afd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.361113 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.361163 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6spw\" (UniqueName: \"kubernetes.io/projected/93d196dc-dc56-4116-bb59-d76601947afd-kube-api-access-x6spw\") on node \"crc\" DevicePath \"\"" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.361180 4810 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.361189 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d196dc-dc56-4116-bb59-d76601947afd-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.450525 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" event={"ID":"93d196dc-dc56-4116-bb59-d76601947afd","Type":"ContainerDied","Data":"2a9c5cc3369984e1f69580a88545a2e1064501722c35a8c871e142d2048a63c3"} Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.450584 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a9c5cc3369984e1f69580a88545a2e1064501722c35a8c871e142d2048a63c3" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.450610 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.549395 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz"] Sep 30 08:27:12 crc kubenswrapper[4810]: E0930 08:27:12.549859 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" containerName="registry-server" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.549880 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" containerName="registry-server" Sep 30 08:27:12 crc kubenswrapper[4810]: E0930 08:27:12.549904 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93d196dc-dc56-4116-bb59-d76601947afd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.549914 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="93d196dc-dc56-4116-bb59-d76601947afd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 08:27:12 crc kubenswrapper[4810]: E0930 08:27:12.549934 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" containerName="extract-content" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.549942 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" containerName="extract-content" Sep 30 08:27:12 crc kubenswrapper[4810]: E0930 08:27:12.549966 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" containerName="extract-utilities" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.549975 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" containerName="extract-utilities" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.550224 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea6b294-b4a3-4e84-a63c-cd3a18a74229" containerName="registry-server" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.550236 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="93d196dc-dc56-4116-bb59-d76601947afd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.551076 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.555006 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.555030 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.556067 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.558445 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.569180 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz"] Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.667824 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vhxfz\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.668066 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vhxfz\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.668428 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt4tw\" (UniqueName: \"kubernetes.io/projected/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-kube-api-access-rt4tw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vhxfz\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.770903 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vhxfz\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.771051 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vhxfz\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.771153 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt4tw\" (UniqueName: \"kubernetes.io/projected/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-kube-api-access-rt4tw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vhxfz\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.779626 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vhxfz\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.790170 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vhxfz\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.800814 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt4tw\" (UniqueName: \"kubernetes.io/projected/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-kube-api-access-rt4tw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vhxfz\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:12 crc kubenswrapper[4810]: I0930 08:27:12.869980 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:13 crc kubenswrapper[4810]: I0930 08:27:13.419703 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz"] Sep 30 08:27:13 crc kubenswrapper[4810]: W0930 08:27:13.421735 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc05f1d1_59b0_45d8_8a52_e0cc456a1207.slice/crio-4939e697ade00773d788ee6aa722dbe702ce05f7988bcf77a02149a8ea9dfe1f WatchSource:0}: Error finding container 4939e697ade00773d788ee6aa722dbe702ce05f7988bcf77a02149a8ea9dfe1f: Status 404 returned error can't find the container with id 4939e697ade00773d788ee6aa722dbe702ce05f7988bcf77a02149a8ea9dfe1f Sep 30 08:27:13 crc kubenswrapper[4810]: I0930 08:27:13.424222 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 08:27:13 crc kubenswrapper[4810]: I0930 08:27:13.463407 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" event={"ID":"dc05f1d1-59b0-45d8-8a52-e0cc456a1207","Type":"ContainerStarted","Data":"4939e697ade00773d788ee6aa722dbe702ce05f7988bcf77a02149a8ea9dfe1f"} Sep 30 08:27:14 crc kubenswrapper[4810]: I0930 08:27:14.476108 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" event={"ID":"dc05f1d1-59b0-45d8-8a52-e0cc456a1207","Type":"ContainerStarted","Data":"80115f80c9946e7e00e61d594a688e3e3a1b4464dc2fcfcaec0d4d53252ada23"} Sep 30 08:27:14 crc kubenswrapper[4810]: I0930 08:27:14.505668 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" podStartSLOduration=1.951197254 podStartE2EDuration="2.505645769s" podCreationTimestamp="2025-09-30 08:27:12 +0000 UTC" firstStartedPulling="2025-09-30 08:27:13.424001076 +0000 UTC m=+1456.876200343" lastFinishedPulling="2025-09-30 08:27:13.978449581 +0000 UTC m=+1457.430648858" observedRunningTime="2025-09-30 08:27:14.497829663 +0000 UTC m=+1457.950028960" watchObservedRunningTime="2025-09-30 08:27:14.505645769 +0000 UTC m=+1457.957845036" Sep 30 08:27:15 crc kubenswrapper[4810]: I0930 08:27:15.911648 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:27:15 crc kubenswrapper[4810]: I0930 08:27:15.911962 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:27:17 crc kubenswrapper[4810]: I0930 08:27:17.508206 4810 generic.go:334] "Generic (PLEG): container finished" podID="dc05f1d1-59b0-45d8-8a52-e0cc456a1207" containerID="80115f80c9946e7e00e61d594a688e3e3a1b4464dc2fcfcaec0d4d53252ada23" exitCode=0 Sep 30 08:27:17 crc kubenswrapper[4810]: I0930 08:27:17.508256 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" event={"ID":"dc05f1d1-59b0-45d8-8a52-e0cc456a1207","Type":"ContainerDied","Data":"80115f80c9946e7e00e61d594a688e3e3a1b4464dc2fcfcaec0d4d53252ada23"} Sep 30 08:27:17 crc kubenswrapper[4810]: I0930 08:27:17.756513 4810 scope.go:117] "RemoveContainer" containerID="9701f0be78a5ff9e61d96211139edfd486723cff8afe5a4c456f626be4594931" Sep 30 08:27:17 crc kubenswrapper[4810]: I0930 08:27:17.803720 4810 scope.go:117] "RemoveContainer" containerID="84151a80e112b5f2857d8d50b1a5e6ff3429de77a28966ba7d97d1da6b61f798" Sep 30 08:27:17 crc kubenswrapper[4810]: I0930 08:27:17.862404 4810 scope.go:117] "RemoveContainer" containerID="ae64ed2fa0a9c068aaa9bf2860a2b815a9b72f9685c9c6814961bf0848668bf1" Sep 30 08:27:18 crc kubenswrapper[4810]: I0930 08:27:18.968695 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.097393 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-inventory\") pod \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.097442 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-ssh-key\") pod \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.097550 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt4tw\" (UniqueName: \"kubernetes.io/projected/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-kube-api-access-rt4tw\") pod \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\" (UID: \"dc05f1d1-59b0-45d8-8a52-e0cc456a1207\") " Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.105045 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-kube-api-access-rt4tw" (OuterVolumeSpecName: "kube-api-access-rt4tw") pod "dc05f1d1-59b0-45d8-8a52-e0cc456a1207" (UID: "dc05f1d1-59b0-45d8-8a52-e0cc456a1207"). InnerVolumeSpecName "kube-api-access-rt4tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.131103 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-inventory" (OuterVolumeSpecName: "inventory") pod "dc05f1d1-59b0-45d8-8a52-e0cc456a1207" (UID: "dc05f1d1-59b0-45d8-8a52-e0cc456a1207"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.140351 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dc05f1d1-59b0-45d8-8a52-e0cc456a1207" (UID: "dc05f1d1-59b0-45d8-8a52-e0cc456a1207"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.200315 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.200380 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.200393 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt4tw\" (UniqueName: \"kubernetes.io/projected/dc05f1d1-59b0-45d8-8a52-e0cc456a1207-kube-api-access-rt4tw\") on node \"crc\" DevicePath \"\"" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.558965 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" event={"ID":"dc05f1d1-59b0-45d8-8a52-e0cc456a1207","Type":"ContainerDied","Data":"4939e697ade00773d788ee6aa722dbe702ce05f7988bcf77a02149a8ea9dfe1f"} Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.559049 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4939e697ade00773d788ee6aa722dbe702ce05f7988bcf77a02149a8ea9dfe1f" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.559137 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vhxfz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.633645 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz"] Sep 30 08:27:19 crc kubenswrapper[4810]: E0930 08:27:19.634485 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc05f1d1-59b0-45d8-8a52-e0cc456a1207" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.634502 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc05f1d1-59b0-45d8-8a52-e0cc456a1207" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.634717 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc05f1d1-59b0-45d8-8a52-e0cc456a1207" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.635410 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.641540 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.641739 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.641837 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.641790 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.664207 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz"] Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.812013 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.812075 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.812100 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.812439 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxtdf\" (UniqueName: \"kubernetes.io/projected/6465dd71-b2e8-4a11-90a8-406ddd763a0c-kube-api-access-dxtdf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.915081 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.915161 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.915201 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.915480 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxtdf\" (UniqueName: \"kubernetes.io/projected/6465dd71-b2e8-4a11-90a8-406ddd763a0c-kube-api-access-dxtdf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.919425 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.920210 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.920792 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.941308 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxtdf\" (UniqueName: \"kubernetes.io/projected/6465dd71-b2e8-4a11-90a8-406ddd763a0c-kube-api-access-dxtdf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:19 crc kubenswrapper[4810]: I0930 08:27:19.957110 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:27:20 crc kubenswrapper[4810]: I0930 08:27:20.598431 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz"] Sep 30 08:27:21 crc kubenswrapper[4810]: I0930 08:27:21.580840 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" event={"ID":"6465dd71-b2e8-4a11-90a8-406ddd763a0c","Type":"ContainerStarted","Data":"b51b6b4a97eb3fd0db81cb055a838103aec6ca77ae983cc90b11e653bdc26b67"} Sep 30 08:27:22 crc kubenswrapper[4810]: I0930 08:27:22.592953 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" event={"ID":"6465dd71-b2e8-4a11-90a8-406ddd763a0c","Type":"ContainerStarted","Data":"b94bd5302fc375ff27f949f309b9c505f07ef06822e96193df18f83558143104"} Sep 30 08:27:22 crc kubenswrapper[4810]: I0930 08:27:22.618119 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" podStartSLOduration=2.85395216 podStartE2EDuration="3.618095215s" podCreationTimestamp="2025-09-30 08:27:19 +0000 UTC" firstStartedPulling="2025-09-30 08:27:20.603753289 +0000 UTC m=+1464.055952546" lastFinishedPulling="2025-09-30 08:27:21.367896324 +0000 UTC m=+1464.820095601" observedRunningTime="2025-09-30 08:27:22.615441478 +0000 UTC m=+1466.067640775" watchObservedRunningTime="2025-09-30 08:27:22.618095215 +0000 UTC m=+1466.070294492" Sep 30 08:27:45 crc kubenswrapper[4810]: I0930 08:27:45.911750 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:27:45 crc kubenswrapper[4810]: I0930 08:27:45.912397 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:27:45 crc kubenswrapper[4810]: I0930 08:27:45.912474 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:27:45 crc kubenswrapper[4810]: I0930 08:27:45.913537 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"56832a708861cd3bbdd73ca09a144eec1b9cab6e8fd4ceb0ce5cd9871bf86c10"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:27:45 crc kubenswrapper[4810]: I0930 08:27:45.913675 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://56832a708861cd3bbdd73ca09a144eec1b9cab6e8fd4ceb0ce5cd9871bf86c10" gracePeriod=600 Sep 30 08:27:46 crc kubenswrapper[4810]: I0930 08:27:46.864699 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="56832a708861cd3bbdd73ca09a144eec1b9cab6e8fd4ceb0ce5cd9871bf86c10" exitCode=0 Sep 30 08:27:46 crc kubenswrapper[4810]: I0930 08:27:46.864998 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"56832a708861cd3bbdd73ca09a144eec1b9cab6e8fd4ceb0ce5cd9871bf86c10"} Sep 30 08:27:46 crc kubenswrapper[4810]: I0930 08:27:46.865599 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54"} Sep 30 08:27:46 crc kubenswrapper[4810]: I0930 08:27:46.865630 4810 scope.go:117] "RemoveContainer" containerID="d77e6c1674e48be39d628cea92dce547143c707ab6bb630e41eb5a3542975559" Sep 30 08:28:18 crc kubenswrapper[4810]: I0930 08:28:18.118104 4810 scope.go:117] "RemoveContainer" containerID="e3e793bcf7de6abdc5f73e4768fbecd1d417ac36ed1653e5ac5026c8ac59b493" Sep 30 08:29:18 crc kubenswrapper[4810]: I0930 08:29:18.236674 4810 scope.go:117] "RemoveContainer" containerID="99541a51d283a02a348a3ddbbbe4d6b66ebcdec80efa338931ab2ff8c1db9866" Sep 30 08:29:18 crc kubenswrapper[4810]: I0930 08:29:18.270916 4810 scope.go:117] "RemoveContainer" containerID="e7e185e2439534208d918105d0d827e8ff9a42a23624373469101ec0b1111e53" Sep 30 08:29:18 crc kubenswrapper[4810]: I0930 08:29:18.295513 4810 scope.go:117] "RemoveContainer" containerID="60451eb9e86ba4154b35b83174bea31f0296f5d0cfeb85eaf93a33f09ba951f1" Sep 30 08:29:18 crc kubenswrapper[4810]: I0930 08:29:18.350080 4810 scope.go:117] "RemoveContainer" containerID="1acf8bfd81f02a2f564cd12a8def41f855578d05ce5835c8c43b9852379cb97c" Sep 30 08:29:18 crc kubenswrapper[4810]: I0930 08:29:18.374491 4810 scope.go:117] "RemoveContainer" containerID="09c06cc5c51068b211fe9ae3b2a140ff729dfb45c98e9055d449f9e357949459" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.172986 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2"] Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.175786 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.177999 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.179416 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.186080 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2"] Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.320990 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-config-volume\") pod \"collect-profiles-29320350-jptm2\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.321447 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-secret-volume\") pod \"collect-profiles-29320350-jptm2\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.321529 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r7rf\" (UniqueName: \"kubernetes.io/projected/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-kube-api-access-5r7rf\") pod \"collect-profiles-29320350-jptm2\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.423935 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r7rf\" (UniqueName: \"kubernetes.io/projected/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-kube-api-access-5r7rf\") pod \"collect-profiles-29320350-jptm2\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.424360 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-config-volume\") pod \"collect-profiles-29320350-jptm2\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.424685 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-secret-volume\") pod \"collect-profiles-29320350-jptm2\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.425294 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-config-volume\") pod \"collect-profiles-29320350-jptm2\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.431620 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-secret-volume\") pod \"collect-profiles-29320350-jptm2\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.444559 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r7rf\" (UniqueName: \"kubernetes.io/projected/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-kube-api-access-5r7rf\") pod \"collect-profiles-29320350-jptm2\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.527646 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:00 crc kubenswrapper[4810]: I0930 08:30:00.999016 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2"] Sep 30 08:30:01 crc kubenswrapper[4810]: I0930 08:30:01.591522 4810 generic.go:334] "Generic (PLEG): container finished" podID="3619e2ef-5db9-4f4a-b8f3-95f93fdba325" containerID="2f23a6139b5411612599b0bd17ef1b314babb37e4d6c38927f7adbec1a400e57" exitCode=0 Sep 30 08:30:01 crc kubenswrapper[4810]: I0930 08:30:01.591574 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" event={"ID":"3619e2ef-5db9-4f4a-b8f3-95f93fdba325","Type":"ContainerDied","Data":"2f23a6139b5411612599b0bd17ef1b314babb37e4d6c38927f7adbec1a400e57"} Sep 30 08:30:01 crc kubenswrapper[4810]: I0930 08:30:01.591602 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" event={"ID":"3619e2ef-5db9-4f4a-b8f3-95f93fdba325","Type":"ContainerStarted","Data":"970d9f170b965c496e012e95f004c7df3a267e689a04dd67313677273a21ced5"} Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.007880 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.183556 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r7rf\" (UniqueName: \"kubernetes.io/projected/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-kube-api-access-5r7rf\") pod \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.184245 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-secret-volume\") pod \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.184634 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-config-volume\") pod \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\" (UID: \"3619e2ef-5db9-4f4a-b8f3-95f93fdba325\") " Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.185345 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-config-volume" (OuterVolumeSpecName: "config-volume") pod "3619e2ef-5db9-4f4a-b8f3-95f93fdba325" (UID: "3619e2ef-5db9-4f4a-b8f3-95f93fdba325"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.185506 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.191583 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3619e2ef-5db9-4f4a-b8f3-95f93fdba325" (UID: "3619e2ef-5db9-4f4a-b8f3-95f93fdba325"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.192639 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-kube-api-access-5r7rf" (OuterVolumeSpecName: "kube-api-access-5r7rf") pod "3619e2ef-5db9-4f4a-b8f3-95f93fdba325" (UID: "3619e2ef-5db9-4f4a-b8f3-95f93fdba325"). InnerVolumeSpecName "kube-api-access-5r7rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.287343 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r7rf\" (UniqueName: \"kubernetes.io/projected/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-kube-api-access-5r7rf\") on node \"crc\" DevicePath \"\"" Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.287390 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3619e2ef-5db9-4f4a-b8f3-95f93fdba325-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.615432 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" event={"ID":"3619e2ef-5db9-4f4a-b8f3-95f93fdba325","Type":"ContainerDied","Data":"970d9f170b965c496e012e95f004c7df3a267e689a04dd67313677273a21ced5"} Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.615497 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="970d9f170b965c496e012e95f004c7df3a267e689a04dd67313677273a21ced5" Sep 30 08:30:03 crc kubenswrapper[4810]: I0930 08:30:03.615553 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2" Sep 30 08:30:15 crc kubenswrapper[4810]: I0930 08:30:15.911438 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:30:15 crc kubenswrapper[4810]: I0930 08:30:15.913650 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:30:18 crc kubenswrapper[4810]: I0930 08:30:18.483602 4810 scope.go:117] "RemoveContainer" containerID="8e9ed83b4714ff39f1383727d350ae37a3b08c7612ca43cc680dcee98117f410" Sep 30 08:30:18 crc kubenswrapper[4810]: I0930 08:30:18.527119 4810 scope.go:117] "RemoveContainer" containerID="fa1405eac25bff82b76214a60c2e7583761f1d6bb64e477c50c3991118f3ed4b" Sep 30 08:30:18 crc kubenswrapper[4810]: I0930 08:30:18.558905 4810 scope.go:117] "RemoveContainer" containerID="5978d931809ac4ce2d5fd8e511370b796f4c1dc65aa700b77eaa0decc93569fd" Sep 30 08:30:18 crc kubenswrapper[4810]: I0930 08:30:18.596910 4810 scope.go:117] "RemoveContainer" containerID="afdfc89354db8734d410ed7e3f0d47c9d0b652c872bc87379955611e3aaa1a50" Sep 30 08:30:18 crc kubenswrapper[4810]: I0930 08:30:18.632495 4810 scope.go:117] "RemoveContainer" containerID="b184a13912d3c57f1ddcad66c8f7bcb029784998b921153d5e8a2672744e65a0" Sep 30 08:30:45 crc kubenswrapper[4810]: I0930 08:30:45.153494 4810 generic.go:334] "Generic (PLEG): container finished" podID="6465dd71-b2e8-4a11-90a8-406ddd763a0c" containerID="b94bd5302fc375ff27f949f309b9c505f07ef06822e96193df18f83558143104" exitCode=0 Sep 30 08:30:45 crc kubenswrapper[4810]: I0930 08:30:45.153674 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" event={"ID":"6465dd71-b2e8-4a11-90a8-406ddd763a0c","Type":"ContainerDied","Data":"b94bd5302fc375ff27f949f309b9c505f07ef06822e96193df18f83558143104"} Sep 30 08:30:45 crc kubenswrapper[4810]: I0930 08:30:45.912014 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:30:45 crc kubenswrapper[4810]: I0930 08:30:45.912103 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.667702 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.787243 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-ssh-key\") pod \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.787389 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-inventory\") pod \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.787513 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-bootstrap-combined-ca-bundle\") pod \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.787751 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxtdf\" (UniqueName: \"kubernetes.io/projected/6465dd71-b2e8-4a11-90a8-406ddd763a0c-kube-api-access-dxtdf\") pod \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\" (UID: \"6465dd71-b2e8-4a11-90a8-406ddd763a0c\") " Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.793640 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6465dd71-b2e8-4a11-90a8-406ddd763a0c" (UID: "6465dd71-b2e8-4a11-90a8-406ddd763a0c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.793802 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6465dd71-b2e8-4a11-90a8-406ddd763a0c-kube-api-access-dxtdf" (OuterVolumeSpecName: "kube-api-access-dxtdf") pod "6465dd71-b2e8-4a11-90a8-406ddd763a0c" (UID: "6465dd71-b2e8-4a11-90a8-406ddd763a0c"). InnerVolumeSpecName "kube-api-access-dxtdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.841422 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6465dd71-b2e8-4a11-90a8-406ddd763a0c" (UID: "6465dd71-b2e8-4a11-90a8-406ddd763a0c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.844149 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-inventory" (OuterVolumeSpecName: "inventory") pod "6465dd71-b2e8-4a11-90a8-406ddd763a0c" (UID: "6465dd71-b2e8-4a11-90a8-406ddd763a0c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.890610 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxtdf\" (UniqueName: \"kubernetes.io/projected/6465dd71-b2e8-4a11-90a8-406ddd763a0c-kube-api-access-dxtdf\") on node \"crc\" DevicePath \"\"" Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.890661 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.890686 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:30:46 crc kubenswrapper[4810]: I0930 08:30:46.890715 4810 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6465dd71-b2e8-4a11-90a8-406ddd763a0c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.184944 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" event={"ID":"6465dd71-b2e8-4a11-90a8-406ddd763a0c","Type":"ContainerDied","Data":"b51b6b4a97eb3fd0db81cb055a838103aec6ca77ae983cc90b11e653bdc26b67"} Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.185003 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b51b6b4a97eb3fd0db81cb055a838103aec6ca77ae983cc90b11e653bdc26b67" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.185046 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.347810 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq"] Sep 30 08:30:47 crc kubenswrapper[4810]: E0930 08:30:47.348257 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6465dd71-b2e8-4a11-90a8-406ddd763a0c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.348295 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6465dd71-b2e8-4a11-90a8-406ddd763a0c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 08:30:47 crc kubenswrapper[4810]: E0930 08:30:47.348331 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3619e2ef-5db9-4f4a-b8f3-95f93fdba325" containerName="collect-profiles" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.348339 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3619e2ef-5db9-4f4a-b8f3-95f93fdba325" containerName="collect-profiles" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.348586 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3619e2ef-5db9-4f4a-b8f3-95f93fdba325" containerName="collect-profiles" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.348625 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6465dd71-b2e8-4a11-90a8-406ddd763a0c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.349391 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.352528 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.352673 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.353008 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.357892 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.368382 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq"] Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.506196 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.506300 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tbhr\" (UniqueName: \"kubernetes.io/projected/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-kube-api-access-2tbhr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.506466 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.608561 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.609533 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tbhr\" (UniqueName: \"kubernetes.io/projected/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-kube-api-access-2tbhr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.609624 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.614048 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.627712 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.633047 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tbhr\" (UniqueName: \"kubernetes.io/projected/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-kube-api-access-2tbhr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:30:47 crc kubenswrapper[4810]: I0930 08:30:47.680659 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:30:48 crc kubenswrapper[4810]: I0930 08:30:48.231234 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq"] Sep 30 08:30:48 crc kubenswrapper[4810]: W0930 08:30:48.238107 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f13a98d_9385_4e17_a8ef_fd0dec1866ac.slice/crio-75d91bb6de7a9b54fedc8f6bcd15b9d19c4a89c07cbc280803a4ba6db5bc96e9 WatchSource:0}: Error finding container 75d91bb6de7a9b54fedc8f6bcd15b9d19c4a89c07cbc280803a4ba6db5bc96e9: Status 404 returned error can't find the container with id 75d91bb6de7a9b54fedc8f6bcd15b9d19c4a89c07cbc280803a4ba6db5bc96e9 Sep 30 08:30:49 crc kubenswrapper[4810]: I0930 08:30:49.210777 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" event={"ID":"3f13a98d-9385-4e17-a8ef-fd0dec1866ac","Type":"ContainerStarted","Data":"75d91bb6de7a9b54fedc8f6bcd15b9d19c4a89c07cbc280803a4ba6db5bc96e9"} Sep 30 08:30:51 crc kubenswrapper[4810]: I0930 08:30:51.246100 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" event={"ID":"3f13a98d-9385-4e17-a8ef-fd0dec1866ac","Type":"ContainerStarted","Data":"3174de3e9fefc7aff8f8c2f613adfddccfa041a6e04e22d0ebfc454488066593"} Sep 30 08:30:51 crc kubenswrapper[4810]: I0930 08:30:51.266164 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" podStartSLOduration=2.573776297 podStartE2EDuration="4.266143617s" podCreationTimestamp="2025-09-30 08:30:47 +0000 UTC" firstStartedPulling="2025-09-30 08:30:48.240445203 +0000 UTC m=+1671.692644480" lastFinishedPulling="2025-09-30 08:30:49.932812503 +0000 UTC m=+1673.385011800" observedRunningTime="2025-09-30 08:30:51.26317352 +0000 UTC m=+1674.715372797" watchObservedRunningTime="2025-09-30 08:30:51.266143617 +0000 UTC m=+1674.718342894" Sep 30 08:30:57 crc kubenswrapper[4810]: I0930 08:30:57.058645 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-h844r"] Sep 30 08:30:57 crc kubenswrapper[4810]: I0930 08:30:57.076744 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-h844r"] Sep 30 08:30:57 crc kubenswrapper[4810]: I0930 08:30:57.324471 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dacf67c-34b5-4c58-9f8d-00adfdcf74b7" path="/var/lib/kubelet/pods/7dacf67c-34b5-4c58-9f8d-00adfdcf74b7/volumes" Sep 30 08:31:01 crc kubenswrapper[4810]: I0930 08:31:01.040684 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-xqn7n"] Sep 30 08:31:01 crc kubenswrapper[4810]: I0930 08:31:01.059596 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-xqn7n"] Sep 30 08:31:01 crc kubenswrapper[4810]: I0930 08:31:01.316957 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="531ec0c6-aa15-4799-b0a7-594fa996f069" path="/var/lib/kubelet/pods/531ec0c6-aa15-4799-b0a7-594fa996f069/volumes" Sep 30 08:31:02 crc kubenswrapper[4810]: I0930 08:31:02.044955 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-tdcfg"] Sep 30 08:31:02 crc kubenswrapper[4810]: I0930 08:31:02.057654 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-tdcfg"] Sep 30 08:31:02 crc kubenswrapper[4810]: I0930 08:31:02.081330 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-95ds9"] Sep 30 08:31:02 crc kubenswrapper[4810]: I0930 08:31:02.093688 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-jgr2f"] Sep 30 08:31:02 crc kubenswrapper[4810]: I0930 08:31:02.104358 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-95ds9"] Sep 30 08:31:02 crc kubenswrapper[4810]: I0930 08:31:02.114460 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-jgr2f"] Sep 30 08:31:03 crc kubenswrapper[4810]: I0930 08:31:03.328774 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="295208bb-eebe-474d-9210-be3a67db28b0" path="/var/lib/kubelet/pods/295208bb-eebe-474d-9210-be3a67db28b0/volumes" Sep 30 08:31:03 crc kubenswrapper[4810]: I0930 08:31:03.330381 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af23fdb8-4a62-427f-8197-ca7a551b7b22" path="/var/lib/kubelet/pods/af23fdb8-4a62-427f-8197-ca7a551b7b22/volumes" Sep 30 08:31:03 crc kubenswrapper[4810]: I0930 08:31:03.333576 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fef53718-de39-49b8-a9af-ef096a5e88ed" path="/var/lib/kubelet/pods/fef53718-de39-49b8-a9af-ef096a5e88ed/volumes" Sep 30 08:31:05 crc kubenswrapper[4810]: I0930 08:31:05.044669 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-c7s2d"] Sep 30 08:31:05 crc kubenswrapper[4810]: I0930 08:31:05.062428 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-c7s2d"] Sep 30 08:31:05 crc kubenswrapper[4810]: I0930 08:31:05.326221 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7af7ec1f-cde8-4a71-953d-ed7426ba7727" path="/var/lib/kubelet/pods/7af7ec1f-cde8-4a71-953d-ed7426ba7727/volumes" Sep 30 08:31:06 crc kubenswrapper[4810]: I0930 08:31:06.040078 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-f9d2p"] Sep 30 08:31:06 crc kubenswrapper[4810]: I0930 08:31:06.059377 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-f9d2p"] Sep 30 08:31:07 crc kubenswrapper[4810]: I0930 08:31:07.037938 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-b9dc-account-create-sxxkk"] Sep 30 08:31:07 crc kubenswrapper[4810]: I0930 08:31:07.049825 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-b9dc-account-create-sxxkk"] Sep 30 08:31:07 crc kubenswrapper[4810]: I0930 08:31:07.327454 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0834a080-3a6e-43c5-95d9-c60c4a10aa66" path="/var/lib/kubelet/pods/0834a080-3a6e-43c5-95d9-c60c4a10aa66/volumes" Sep 30 08:31:07 crc kubenswrapper[4810]: I0930 08:31:07.329103 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ab68570-8ed3-4c73-94da-88ced0b8d75e" path="/var/lib/kubelet/pods/7ab68570-8ed3-4c73-94da-88ced0b8d75e/volumes" Sep 30 08:31:11 crc kubenswrapper[4810]: I0930 08:31:11.045650 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-dfa1-account-create-tnx8l"] Sep 30 08:31:11 crc kubenswrapper[4810]: I0930 08:31:11.061919 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-15e4-account-create-ds56p"] Sep 30 08:31:11 crc kubenswrapper[4810]: I0930 08:31:11.073689 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-588d-account-create-hl9zx"] Sep 30 08:31:11 crc kubenswrapper[4810]: I0930 08:31:11.083742 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-dfa1-account-create-tnx8l"] Sep 30 08:31:11 crc kubenswrapper[4810]: I0930 08:31:11.090967 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-588d-account-create-hl9zx"] Sep 30 08:31:11 crc kubenswrapper[4810]: I0930 08:31:11.098043 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-15e4-account-create-ds56p"] Sep 30 08:31:11 crc kubenswrapper[4810]: I0930 08:31:11.321470 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="127b94a4-4969-4a8d-b708-1b1c6f61d6af" path="/var/lib/kubelet/pods/127b94a4-4969-4a8d-b708-1b1c6f61d6af/volumes" Sep 30 08:31:11 crc kubenswrapper[4810]: I0930 08:31:11.322751 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d505f4a8-0fe5-4cc5-9e55-369f42edd90a" path="/var/lib/kubelet/pods/d505f4a8-0fe5-4cc5-9e55-369f42edd90a/volumes" Sep 30 08:31:11 crc kubenswrapper[4810]: I0930 08:31:11.323537 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcd581d7-6a83-4822-822b-421633417bdf" path="/var/lib/kubelet/pods/fcd581d7-6a83-4822-822b-421633417bdf/volumes" Sep 30 08:31:12 crc kubenswrapper[4810]: I0930 08:31:12.062559 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-410c-account-create-4czqf"] Sep 30 08:31:12 crc kubenswrapper[4810]: I0930 08:31:12.083976 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-410c-account-create-4czqf"] Sep 30 08:31:13 crc kubenswrapper[4810]: I0930 08:31:13.329006 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca25948f-0522-43b1-9b6b-a7ce98207be0" path="/var/lib/kubelet/pods/ca25948f-0522-43b1-9b6b-a7ce98207be0/volumes" Sep 30 08:31:15 crc kubenswrapper[4810]: I0930 08:31:15.911376 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:31:15 crc kubenswrapper[4810]: I0930 08:31:15.913360 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:31:15 crc kubenswrapper[4810]: I0930 08:31:15.913588 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:31:15 crc kubenswrapper[4810]: I0930 08:31:15.914986 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:31:15 crc kubenswrapper[4810]: I0930 08:31:15.915331 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" gracePeriod=600 Sep 30 08:31:16 crc kubenswrapper[4810]: E0930 08:31:16.066897 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:31:16 crc kubenswrapper[4810]: I0930 08:31:16.564408 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" exitCode=0 Sep 30 08:31:16 crc kubenswrapper[4810]: I0930 08:31:16.564494 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54"} Sep 30 08:31:16 crc kubenswrapper[4810]: I0930 08:31:16.564749 4810 scope.go:117] "RemoveContainer" containerID="56832a708861cd3bbdd73ca09a144eec1b9cab6e8fd4ceb0ce5cd9871bf86c10" Sep 30 08:31:16 crc kubenswrapper[4810]: I0930 08:31:16.565298 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:31:16 crc kubenswrapper[4810]: E0930 08:31:16.565674 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:31:17 crc kubenswrapper[4810]: I0930 08:31:17.063215 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5fa8-account-create-xtqgd"] Sep 30 08:31:17 crc kubenswrapper[4810]: I0930 08:31:17.074159 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5fa8-account-create-xtqgd"] Sep 30 08:31:17 crc kubenswrapper[4810]: I0930 08:31:17.323164 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="737b1bf9-6bc6-49a7-8913-5c78af3f1e82" path="/var/lib/kubelet/pods/737b1bf9-6bc6-49a7-8913-5c78af3f1e82/volumes" Sep 30 08:31:18 crc kubenswrapper[4810]: I0930 08:31:18.704881 4810 scope.go:117] "RemoveContainer" containerID="e1004c44a710c6b1586c7ec58d8cb79dfbbe1bc7012bef84028057c07d6549a6" Sep 30 08:31:18 crc kubenswrapper[4810]: I0930 08:31:18.739163 4810 scope.go:117] "RemoveContainer" containerID="c9cf8eb6b3fd5cc334ba71d12ecc662c9cc0c7d77bf3a629147481625c1bf1d2" Sep 30 08:31:18 crc kubenswrapper[4810]: I0930 08:31:18.769557 4810 scope.go:117] "RemoveContainer" containerID="3d34db2a649ce680a9f15f98539d7a3bf4ba3020b2242e6b1eb8a4bec6e3bf54" Sep 30 08:31:18 crc kubenswrapper[4810]: I0930 08:31:18.814271 4810 scope.go:117] "RemoveContainer" containerID="065d5e95f2e03734ff22ed881a330c63409087875ce7e2d6d077ed8c07c48840" Sep 30 08:31:18 crc kubenswrapper[4810]: I0930 08:31:18.852571 4810 scope.go:117] "RemoveContainer" containerID="508caefa1a2537eeb65f5e2cbea3da41cd51ebc16c43bf2a167622e5f83972b0" Sep 30 08:31:18 crc kubenswrapper[4810]: I0930 08:31:18.906966 4810 scope.go:117] "RemoveContainer" containerID="3c0e2d67e5392c64e8647ef5661596e0c743710de0f86368f60b95241e617e0a" Sep 30 08:31:18 crc kubenswrapper[4810]: I0930 08:31:18.940473 4810 scope.go:117] "RemoveContainer" containerID="47ab22daa69a39d7b3c6e9d6184d773f0f9535293c3381650bb88a69e4e68d81" Sep 30 08:31:18 crc kubenswrapper[4810]: I0930 08:31:18.985494 4810 scope.go:117] "RemoveContainer" containerID="ed4ab215943a49af7f1431e9a25e5b252f7be9e0092b220237217d4ba4095e24" Sep 30 08:31:19 crc kubenswrapper[4810]: I0930 08:31:19.010797 4810 scope.go:117] "RemoveContainer" containerID="b2e64049309fffd0cc128b6123caa07c1ccb1191f677273a1715030c3141fdbd" Sep 30 08:31:19 crc kubenswrapper[4810]: I0930 08:31:19.033467 4810 scope.go:117] "RemoveContainer" containerID="64ac1a88b3cc20b7584a4718f6cea16810f7a2767bcba16b6844af74548d9def" Sep 30 08:31:19 crc kubenswrapper[4810]: I0930 08:31:19.055754 4810 scope.go:117] "RemoveContainer" containerID="bd85fffad441a4190e82453ea8a6577c502fee4b917a4cb4f2ebac02d71c9e73" Sep 30 08:31:19 crc kubenswrapper[4810]: I0930 08:31:19.080044 4810 scope.go:117] "RemoveContainer" containerID="7efd85c00404d4b93ee15bfe8912533c23ce84d7536d2e7ee22fde649024d43e" Sep 30 08:31:19 crc kubenswrapper[4810]: I0930 08:31:19.102220 4810 scope.go:117] "RemoveContainer" containerID="5fa449acb9ced2efab045b346eb0ccd4d51e48bd2b87ceab7395768c4f9460dd" Sep 30 08:31:19 crc kubenswrapper[4810]: I0930 08:31:19.126434 4810 scope.go:117] "RemoveContainer" containerID="d5506f9d5f727624787519b93d0b7fd28688f63ac436751c0bb78b6ce632b1b1" Sep 30 08:31:19 crc kubenswrapper[4810]: I0930 08:31:19.150938 4810 scope.go:117] "RemoveContainer" containerID="65ba002271130d507ec502565bd2b4c5eb79b9fc83389096ba418751bce7be6a" Sep 30 08:31:20 crc kubenswrapper[4810]: I0930 08:31:20.040822 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c962-account-create-swr7n"] Sep 30 08:31:20 crc kubenswrapper[4810]: I0930 08:31:20.058738 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c962-account-create-swr7n"] Sep 30 08:31:21 crc kubenswrapper[4810]: I0930 08:31:21.317499 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8962a403-5d8b-4295-8c59-438ccf65fdbc" path="/var/lib/kubelet/pods/8962a403-5d8b-4295-8c59-438ccf65fdbc/volumes" Sep 30 08:31:24 crc kubenswrapper[4810]: I0930 08:31:24.041303 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-2gdsp"] Sep 30 08:31:24 crc kubenswrapper[4810]: I0930 08:31:24.057014 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-2gdsp"] Sep 30 08:31:25 crc kubenswrapper[4810]: I0930 08:31:25.329331 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24dc5231-a128-46dc-b0a2-5d13524ec5db" path="/var/lib/kubelet/pods/24dc5231-a128-46dc-b0a2-5d13524ec5db/volumes" Sep 30 08:31:29 crc kubenswrapper[4810]: I0930 08:31:29.307448 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:31:29 crc kubenswrapper[4810]: E0930 08:31:29.308924 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:31:41 crc kubenswrapper[4810]: I0930 08:31:41.306875 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:31:41 crc kubenswrapper[4810]: E0930 08:31:41.307757 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:31:42 crc kubenswrapper[4810]: I0930 08:31:42.053867 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xd7mq"] Sep 30 08:31:42 crc kubenswrapper[4810]: I0930 08:31:42.069187 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xd7mq"] Sep 30 08:31:43 crc kubenswrapper[4810]: I0930 08:31:43.330918 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df504df8-0ce9-4250-9ffc-db3f3e1ee26e" path="/var/lib/kubelet/pods/df504df8-0ce9-4250-9ffc-db3f3e1ee26e/volumes" Sep 30 08:31:54 crc kubenswrapper[4810]: I0930 08:31:54.306728 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:31:54 crc kubenswrapper[4810]: E0930 08:31:54.307726 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:32:09 crc kubenswrapper[4810]: I0930 08:32:09.306789 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:32:09 crc kubenswrapper[4810]: E0930 08:32:09.307910 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:32:16 crc kubenswrapper[4810]: I0930 08:32:16.059322 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-d2fzh"] Sep 30 08:32:16 crc kubenswrapper[4810]: I0930 08:32:16.074115 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-d2fzh"] Sep 30 08:32:17 crc kubenswrapper[4810]: I0930 08:32:17.329560 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="007f0b25-94b5-4529-b835-08cdc34e0d6e" path="/var/lib/kubelet/pods/007f0b25-94b5-4529-b835-08cdc34e0d6e/volumes" Sep 30 08:32:19 crc kubenswrapper[4810]: I0930 08:32:19.407835 4810 scope.go:117] "RemoveContainer" containerID="ef0de82656893412cdb8c0deb6734b4b0c1bf728448dcd0e3a89874ae3bba4c7" Sep 30 08:32:19 crc kubenswrapper[4810]: I0930 08:32:19.436366 4810 scope.go:117] "RemoveContainer" containerID="03579d2144c26fe931742e864142471db253a179541dd1e4eea42c5429bab5b8" Sep 30 08:32:19 crc kubenswrapper[4810]: I0930 08:32:19.510547 4810 scope.go:117] "RemoveContainer" containerID="1fccb98ff7343b898c2ff0e27bbf9f156471d551d329b7938b1a77fd54308054" Sep 30 08:32:19 crc kubenswrapper[4810]: I0930 08:32:19.545906 4810 scope.go:117] "RemoveContainer" containerID="b0373a01d538c3646517d116acc2f25bed8ebfbd8ad9eb70be0f142464ef077b" Sep 30 08:32:19 crc kubenswrapper[4810]: I0930 08:32:19.587402 4810 scope.go:117] "RemoveContainer" containerID="96a88818d3c567e4575fc2096f67461bfab333bbf939f7fa14727864172c5f4f" Sep 30 08:32:19 crc kubenswrapper[4810]: I0930 08:32:19.662880 4810 scope.go:117] "RemoveContainer" containerID="e81c0d5ff5a09a1428e069d6c9847bd00a288bad6b4ac403b71e606e22914f9e" Sep 30 08:32:23 crc kubenswrapper[4810]: I0930 08:32:23.064447 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-8xkhj"] Sep 30 08:32:23 crc kubenswrapper[4810]: I0930 08:32:23.075372 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-8xkhj"] Sep 30 08:32:23 crc kubenswrapper[4810]: I0930 08:32:23.306885 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:32:23 crc kubenswrapper[4810]: E0930 08:32:23.307461 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:32:23 crc kubenswrapper[4810]: I0930 08:32:23.318843 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f" path="/var/lib/kubelet/pods/f9a3f8dd-7e51-445d-bb0d-a2df3bd6802f/volumes" Sep 30 08:32:25 crc kubenswrapper[4810]: I0930 08:32:25.050144 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-x57vl"] Sep 30 08:32:25 crc kubenswrapper[4810]: I0930 08:32:25.072686 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-x57vl"] Sep 30 08:32:25 crc kubenswrapper[4810]: I0930 08:32:25.320054 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65f3abf6-0015-41fc-9a40-0ed346bb84ec" path="/var/lib/kubelet/pods/65f3abf6-0015-41fc-9a40-0ed346bb84ec/volumes" Sep 30 08:32:32 crc kubenswrapper[4810]: I0930 08:32:32.478340 4810 generic.go:334] "Generic (PLEG): container finished" podID="3f13a98d-9385-4e17-a8ef-fd0dec1866ac" containerID="3174de3e9fefc7aff8f8c2f613adfddccfa041a6e04e22d0ebfc454488066593" exitCode=0 Sep 30 08:32:32 crc kubenswrapper[4810]: I0930 08:32:32.478460 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" event={"ID":"3f13a98d-9385-4e17-a8ef-fd0dec1866ac","Type":"ContainerDied","Data":"3174de3e9fefc7aff8f8c2f613adfddccfa041a6e04e22d0ebfc454488066593"} Sep 30 08:32:33 crc kubenswrapper[4810]: I0930 08:32:33.955841 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.065247 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tbhr\" (UniqueName: \"kubernetes.io/projected/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-kube-api-access-2tbhr\") pod \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.065704 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-inventory\") pod \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.065959 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-ssh-key\") pod \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\" (UID: \"3f13a98d-9385-4e17-a8ef-fd0dec1866ac\") " Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.071131 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-kube-api-access-2tbhr" (OuterVolumeSpecName: "kube-api-access-2tbhr") pod "3f13a98d-9385-4e17-a8ef-fd0dec1866ac" (UID: "3f13a98d-9385-4e17-a8ef-fd0dec1866ac"). InnerVolumeSpecName "kube-api-access-2tbhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.096020 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-inventory" (OuterVolumeSpecName: "inventory") pod "3f13a98d-9385-4e17-a8ef-fd0dec1866ac" (UID: "3f13a98d-9385-4e17-a8ef-fd0dec1866ac"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.096212 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3f13a98d-9385-4e17-a8ef-fd0dec1866ac" (UID: "3f13a98d-9385-4e17-a8ef-fd0dec1866ac"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.169217 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.169529 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.169543 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tbhr\" (UniqueName: \"kubernetes.io/projected/3f13a98d-9385-4e17-a8ef-fd0dec1866ac-kube-api-access-2tbhr\") on node \"crc\" DevicePath \"\"" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.508318 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" event={"ID":"3f13a98d-9385-4e17-a8ef-fd0dec1866ac","Type":"ContainerDied","Data":"75d91bb6de7a9b54fedc8f6bcd15b9d19c4a89c07cbc280803a4ba6db5bc96e9"} Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.508391 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75d91bb6de7a9b54fedc8f6bcd15b9d19c4a89c07cbc280803a4ba6db5bc96e9" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.508564 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.690665 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4"] Sep 30 08:32:34 crc kubenswrapper[4810]: E0930 08:32:34.691139 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f13a98d-9385-4e17-a8ef-fd0dec1866ac" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.691159 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f13a98d-9385-4e17-a8ef-fd0dec1866ac" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.691375 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f13a98d-9385-4e17-a8ef-fd0dec1866ac" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.692042 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.694249 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.694449 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.694585 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.694745 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.716885 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4"] Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.882608 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s67br\" (UniqueName: \"kubernetes.io/projected/459ff4e4-7043-4e6e-9774-97dab472459c-kube-api-access-s67br\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22dr4\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.882668 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22dr4\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.882715 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22dr4\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.984560 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s67br\" (UniqueName: \"kubernetes.io/projected/459ff4e4-7043-4e6e-9774-97dab472459c-kube-api-access-s67br\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22dr4\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.984639 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22dr4\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.984687 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22dr4\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.989347 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22dr4\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:32:34 crc kubenswrapper[4810]: I0930 08:32:34.989936 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22dr4\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:32:35 crc kubenswrapper[4810]: I0930 08:32:35.013697 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s67br\" (UniqueName: \"kubernetes.io/projected/459ff4e4-7043-4e6e-9774-97dab472459c-kube-api-access-s67br\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-22dr4\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:32:35 crc kubenswrapper[4810]: I0930 08:32:35.309144 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:32:35 crc kubenswrapper[4810]: I0930 08:32:35.945898 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4"] Sep 30 08:32:35 crc kubenswrapper[4810]: I0930 08:32:35.950141 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 08:32:36 crc kubenswrapper[4810]: I0930 08:32:36.534968 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" event={"ID":"459ff4e4-7043-4e6e-9774-97dab472459c","Type":"ContainerStarted","Data":"2d147159d5555131f8197c324e8d35537633497a7753ef1f98f93145266ecfa7"} Sep 30 08:32:37 crc kubenswrapper[4810]: I0930 08:32:37.354813 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:32:37 crc kubenswrapper[4810]: E0930 08:32:37.355358 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:32:37 crc kubenswrapper[4810]: I0930 08:32:37.550100 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" event={"ID":"459ff4e4-7043-4e6e-9774-97dab472459c","Type":"ContainerStarted","Data":"036576b35267c41ffbde8c1c0ec2307e5a02460dc801e33204baabcb410d79c7"} Sep 30 08:32:37 crc kubenswrapper[4810]: I0930 08:32:37.578699 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" podStartSLOduration=2.990612198 podStartE2EDuration="3.578638542s" podCreationTimestamp="2025-09-30 08:32:34 +0000 UTC" firstStartedPulling="2025-09-30 08:32:35.949847098 +0000 UTC m=+1779.402046365" lastFinishedPulling="2025-09-30 08:32:36.537873432 +0000 UTC m=+1779.990072709" observedRunningTime="2025-09-30 08:32:37.574797491 +0000 UTC m=+1781.026996758" watchObservedRunningTime="2025-09-30 08:32:37.578638542 +0000 UTC m=+1781.030837829" Sep 30 08:32:39 crc kubenswrapper[4810]: I0930 08:32:39.045735 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4c6qh"] Sep 30 08:32:39 crc kubenswrapper[4810]: I0930 08:32:39.058717 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4c6qh"] Sep 30 08:32:39 crc kubenswrapper[4810]: I0930 08:32:39.321193 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff6ce927-f393-4895-a7b6-af266138c8c1" path="/var/lib/kubelet/pods/ff6ce927-f393-4895-a7b6-af266138c8c1/volumes" Sep 30 08:32:45 crc kubenswrapper[4810]: I0930 08:32:45.038841 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-k5dlb"] Sep 30 08:32:45 crc kubenswrapper[4810]: I0930 08:32:45.050548 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-k5dlb"] Sep 30 08:32:45 crc kubenswrapper[4810]: I0930 08:32:45.317843 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="913aa7a2-f310-4b1e-a403-2e5f98b339b5" path="/var/lib/kubelet/pods/913aa7a2-f310-4b1e-a403-2e5f98b339b5/volumes" Sep 30 08:32:47 crc kubenswrapper[4810]: I0930 08:32:47.054514 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-z7k2l"] Sep 30 08:32:47 crc kubenswrapper[4810]: I0930 08:32:47.065412 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-z7k2l"] Sep 30 08:32:47 crc kubenswrapper[4810]: I0930 08:32:47.325033 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e199d2a8-2a58-47c7-8018-13b29c37c2ad" path="/var/lib/kubelet/pods/e199d2a8-2a58-47c7-8018-13b29c37c2ad/volumes" Sep 30 08:32:50 crc kubenswrapper[4810]: I0930 08:32:50.307482 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:32:50 crc kubenswrapper[4810]: E0930 08:32:50.308333 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:33:05 crc kubenswrapper[4810]: I0930 08:33:05.307236 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:33:05 crc kubenswrapper[4810]: E0930 08:33:05.308351 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:33:19 crc kubenswrapper[4810]: I0930 08:33:19.307056 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:33:19 crc kubenswrapper[4810]: E0930 08:33:19.307975 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:33:19 crc kubenswrapper[4810]: I0930 08:33:19.888412 4810 scope.go:117] "RemoveContainer" containerID="38f6fd6436640fd7a28907681d9e1a4ae351640a40307e3213e3af643dfa869e" Sep 30 08:33:19 crc kubenswrapper[4810]: I0930 08:33:19.937632 4810 scope.go:117] "RemoveContainer" containerID="60c921d9ce7b936f7c16b702afca10ea57bac1a8f2e2198e37531bdc3102bc8c" Sep 30 08:33:20 crc kubenswrapper[4810]: I0930 08:33:20.016017 4810 scope.go:117] "RemoveContainer" containerID="a8483f9453c3170461fd7f633d48cb345a6e19d91c2506f88094b43631ddb18c" Sep 30 08:33:20 crc kubenswrapper[4810]: I0930 08:33:20.071602 4810 scope.go:117] "RemoveContainer" containerID="2096808c93bcdbf171afae758a4667a0e204f744f187817e722454267706b358" Sep 30 08:33:20 crc kubenswrapper[4810]: I0930 08:33:20.139152 4810 scope.go:117] "RemoveContainer" containerID="f0abbb76c1572c27e8547117cdc22146c0ce28960bd84e36efdb24d3e672034c" Sep 30 08:33:31 crc kubenswrapper[4810]: I0930 08:33:31.307478 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:33:31 crc kubenswrapper[4810]: E0930 08:33:31.308350 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:33:36 crc kubenswrapper[4810]: I0930 08:33:36.059551 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-nnzr4"] Sep 30 08:33:36 crc kubenswrapper[4810]: I0930 08:33:36.079217 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-xjw8l"] Sep 30 08:33:36 crc kubenswrapper[4810]: I0930 08:33:36.099809 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8q22m"] Sep 30 08:33:36 crc kubenswrapper[4810]: I0930 08:33:36.107070 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-xjw8l"] Sep 30 08:33:36 crc kubenswrapper[4810]: I0930 08:33:36.114831 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-nnzr4"] Sep 30 08:33:36 crc kubenswrapper[4810]: I0930 08:33:36.122193 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8q22m"] Sep 30 08:33:37 crc kubenswrapper[4810]: I0930 08:33:37.327411 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34b10dbc-a43d-4398-a60a-86c1c8101e8a" path="/var/lib/kubelet/pods/34b10dbc-a43d-4398-a60a-86c1c8101e8a/volumes" Sep 30 08:33:37 crc kubenswrapper[4810]: I0930 08:33:37.328929 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c40b4429-5a41-4ea8-b9ac-17bba8788f6a" path="/var/lib/kubelet/pods/c40b4429-5a41-4ea8-b9ac-17bba8788f6a/volumes" Sep 30 08:33:37 crc kubenswrapper[4810]: I0930 08:33:37.329979 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d61500fa-1a5e-420a-a0d9-990c98754b13" path="/var/lib/kubelet/pods/d61500fa-1a5e-420a-a0d9-990c98754b13/volumes" Sep 30 08:33:43 crc kubenswrapper[4810]: I0930 08:33:43.307412 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:33:43 crc kubenswrapper[4810]: E0930 08:33:43.308427 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:33:46 crc kubenswrapper[4810]: I0930 08:33:46.051909 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-00fe-account-create-6v4sg"] Sep 30 08:33:46 crc kubenswrapper[4810]: I0930 08:33:46.071718 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c663-account-create-hx8d6"] Sep 30 08:33:46 crc kubenswrapper[4810]: I0930 08:33:46.087430 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-00fe-account-create-6v4sg"] Sep 30 08:33:46 crc kubenswrapper[4810]: I0930 08:33:46.096656 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c663-account-create-hx8d6"] Sep 30 08:33:46 crc kubenswrapper[4810]: I0930 08:33:46.107423 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-4b13-account-create-t2xtd"] Sep 30 08:33:46 crc kubenswrapper[4810]: I0930 08:33:46.117211 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-4b13-account-create-t2xtd"] Sep 30 08:33:47 crc kubenswrapper[4810]: I0930 08:33:47.327648 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="608ee485-d5c0-4724-9101-54958f7344b0" path="/var/lib/kubelet/pods/608ee485-d5c0-4724-9101-54958f7344b0/volumes" Sep 30 08:33:47 crc kubenswrapper[4810]: I0930 08:33:47.329308 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="913c4569-9c55-4ba3-8a28-b4a23f294852" path="/var/lib/kubelet/pods/913c4569-9c55-4ba3-8a28-b4a23f294852/volumes" Sep 30 08:33:47 crc kubenswrapper[4810]: I0930 08:33:47.330316 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb" path="/var/lib/kubelet/pods/a5e8c4d2-74a5-4bc3-9b5c-44399ac652fb/volumes" Sep 30 08:33:56 crc kubenswrapper[4810]: I0930 08:33:56.308450 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:33:56 crc kubenswrapper[4810]: E0930 08:33:56.309751 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:33:56 crc kubenswrapper[4810]: I0930 08:33:56.465473 4810 generic.go:334] "Generic (PLEG): container finished" podID="459ff4e4-7043-4e6e-9774-97dab472459c" containerID="036576b35267c41ffbde8c1c0ec2307e5a02460dc801e33204baabcb410d79c7" exitCode=0 Sep 30 08:33:56 crc kubenswrapper[4810]: I0930 08:33:56.465528 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" event={"ID":"459ff4e4-7043-4e6e-9774-97dab472459c","Type":"ContainerDied","Data":"036576b35267c41ffbde8c1c0ec2307e5a02460dc801e33204baabcb410d79c7"} Sep 30 08:33:57 crc kubenswrapper[4810]: I0930 08:33:57.971435 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.091125 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-inventory\") pod \"459ff4e4-7043-4e6e-9774-97dab472459c\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.091189 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-ssh-key\") pod \"459ff4e4-7043-4e6e-9774-97dab472459c\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.091425 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s67br\" (UniqueName: \"kubernetes.io/projected/459ff4e4-7043-4e6e-9774-97dab472459c-kube-api-access-s67br\") pod \"459ff4e4-7043-4e6e-9774-97dab472459c\" (UID: \"459ff4e4-7043-4e6e-9774-97dab472459c\") " Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.096848 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/459ff4e4-7043-4e6e-9774-97dab472459c-kube-api-access-s67br" (OuterVolumeSpecName: "kube-api-access-s67br") pod "459ff4e4-7043-4e6e-9774-97dab472459c" (UID: "459ff4e4-7043-4e6e-9774-97dab472459c"). InnerVolumeSpecName "kube-api-access-s67br". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.120093 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "459ff4e4-7043-4e6e-9774-97dab472459c" (UID: "459ff4e4-7043-4e6e-9774-97dab472459c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.129923 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-inventory" (OuterVolumeSpecName: "inventory") pod "459ff4e4-7043-4e6e-9774-97dab472459c" (UID: "459ff4e4-7043-4e6e-9774-97dab472459c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.195216 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s67br\" (UniqueName: \"kubernetes.io/projected/459ff4e4-7043-4e6e-9774-97dab472459c-kube-api-access-s67br\") on node \"crc\" DevicePath \"\"" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.195300 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.195314 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/459ff4e4-7043-4e6e-9774-97dab472459c-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.495208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" event={"ID":"459ff4e4-7043-4e6e-9774-97dab472459c","Type":"ContainerDied","Data":"2d147159d5555131f8197c324e8d35537633497a7753ef1f98f93145266ecfa7"} Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.495311 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d147159d5555131f8197c324e8d35537633497a7753ef1f98f93145266ecfa7" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.495363 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-22dr4" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.597473 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b"] Sep 30 08:33:58 crc kubenswrapper[4810]: E0930 08:33:58.597971 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459ff4e4-7043-4e6e-9774-97dab472459c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.597993 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="459ff4e4-7043-4e6e-9774-97dab472459c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.598258 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="459ff4e4-7043-4e6e-9774-97dab472459c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.599075 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.602392 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.602819 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.602955 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.603744 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.610705 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tvq2\" (UniqueName: \"kubernetes.io/projected/a86849c2-c174-4b69-84a0-281bb1afda6d-kube-api-access-7tvq2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.610772 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.610875 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.627198 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b"] Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.713526 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tvq2\" (UniqueName: \"kubernetes.io/projected/a86849c2-c174-4b69-84a0-281bb1afda6d-kube-api-access-7tvq2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.713591 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.713677 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.718758 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.722327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.737786 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tvq2\" (UniqueName: \"kubernetes.io/projected/a86849c2-c174-4b69-84a0-281bb1afda6d-kube-api-access-7tvq2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:33:58 crc kubenswrapper[4810]: I0930 08:33:58.921950 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:33:59 crc kubenswrapper[4810]: I0930 08:33:59.339342 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b"] Sep 30 08:33:59 crc kubenswrapper[4810]: I0930 08:33:59.509418 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" event={"ID":"a86849c2-c174-4b69-84a0-281bb1afda6d","Type":"ContainerStarted","Data":"a2449ba74c0378ad982495463b7d21a1ae041375e115b4b7804c04acb5bb477a"} Sep 30 08:34:00 crc kubenswrapper[4810]: I0930 08:34:00.523610 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" event={"ID":"a86849c2-c174-4b69-84a0-281bb1afda6d","Type":"ContainerStarted","Data":"52314b0d530f5ceb8402e6480ea9dc09d5a7018a98e98877d7cded0fb6ed4a87"} Sep 30 08:34:05 crc kubenswrapper[4810]: I0930 08:34:05.602761 4810 generic.go:334] "Generic (PLEG): container finished" podID="a86849c2-c174-4b69-84a0-281bb1afda6d" containerID="52314b0d530f5ceb8402e6480ea9dc09d5a7018a98e98877d7cded0fb6ed4a87" exitCode=0 Sep 30 08:34:05 crc kubenswrapper[4810]: I0930 08:34:05.602879 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" event={"ID":"a86849c2-c174-4b69-84a0-281bb1afda6d","Type":"ContainerDied","Data":"52314b0d530f5ceb8402e6480ea9dc09d5a7018a98e98877d7cded0fb6ed4a87"} Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.101359 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.218669 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tvq2\" (UniqueName: \"kubernetes.io/projected/a86849c2-c174-4b69-84a0-281bb1afda6d-kube-api-access-7tvq2\") pod \"a86849c2-c174-4b69-84a0-281bb1afda6d\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.218788 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-ssh-key\") pod \"a86849c2-c174-4b69-84a0-281bb1afda6d\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.218828 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-inventory\") pod \"a86849c2-c174-4b69-84a0-281bb1afda6d\" (UID: \"a86849c2-c174-4b69-84a0-281bb1afda6d\") " Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.226262 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a86849c2-c174-4b69-84a0-281bb1afda6d-kube-api-access-7tvq2" (OuterVolumeSpecName: "kube-api-access-7tvq2") pod "a86849c2-c174-4b69-84a0-281bb1afda6d" (UID: "a86849c2-c174-4b69-84a0-281bb1afda6d"). InnerVolumeSpecName "kube-api-access-7tvq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.259194 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-inventory" (OuterVolumeSpecName: "inventory") pod "a86849c2-c174-4b69-84a0-281bb1afda6d" (UID: "a86849c2-c174-4b69-84a0-281bb1afda6d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.274627 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a86849c2-c174-4b69-84a0-281bb1afda6d" (UID: "a86849c2-c174-4b69-84a0-281bb1afda6d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.324362 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tvq2\" (UniqueName: \"kubernetes.io/projected/a86849c2-c174-4b69-84a0-281bb1afda6d-kube-api-access-7tvq2\") on node \"crc\" DevicePath \"\"" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.324622 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.324702 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a86849c2-c174-4b69-84a0-281bb1afda6d-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.623714 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" event={"ID":"a86849c2-c174-4b69-84a0-281bb1afda6d","Type":"ContainerDied","Data":"a2449ba74c0378ad982495463b7d21a1ae041375e115b4b7804c04acb5bb477a"} Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.623755 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2449ba74c0378ad982495463b7d21a1ae041375e115b4b7804c04acb5bb477a" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.623776 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.726568 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7"] Sep 30 08:34:07 crc kubenswrapper[4810]: E0930 08:34:07.727168 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86849c2-c174-4b69-84a0-281bb1afda6d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.727302 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86849c2-c174-4b69-84a0-281bb1afda6d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.727591 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a86849c2-c174-4b69-84a0-281bb1afda6d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.728298 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.731434 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.731571 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.731573 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.731729 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.791335 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7"] Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.836020 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9gnn7\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.836128 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9gnn7\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.836230 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwkch\" (UniqueName: \"kubernetes.io/projected/75b17470-760f-4fa6-9550-c2281ed5ae6b-kube-api-access-pwkch\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9gnn7\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.938114 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwkch\" (UniqueName: \"kubernetes.io/projected/75b17470-760f-4fa6-9550-c2281ed5ae6b-kube-api-access-pwkch\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9gnn7\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.938561 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9gnn7\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.938630 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9gnn7\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.947719 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9gnn7\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.953230 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9gnn7\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:07 crc kubenswrapper[4810]: I0930 08:34:07.971207 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwkch\" (UniqueName: \"kubernetes.io/projected/75b17470-760f-4fa6-9550-c2281ed5ae6b-kube-api-access-pwkch\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9gnn7\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:08 crc kubenswrapper[4810]: I0930 08:34:08.043620 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:08 crc kubenswrapper[4810]: I0930 08:34:08.614799 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7"] Sep 30 08:34:08 crc kubenswrapper[4810]: I0930 08:34:08.635574 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" event={"ID":"75b17470-760f-4fa6-9550-c2281ed5ae6b","Type":"ContainerStarted","Data":"40eeb9b78dbc3bff2babf28609ba8f31deaff7030018862b0f5d4d026cbf11c8"} Sep 30 08:34:09 crc kubenswrapper[4810]: I0930 08:34:09.652883 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" event={"ID":"75b17470-760f-4fa6-9550-c2281ed5ae6b","Type":"ContainerStarted","Data":"fbba86a400be745516633e58871e85f3975e6f145b24825d72c60b684d992017"} Sep 30 08:34:09 crc kubenswrapper[4810]: I0930 08:34:09.684219 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" podStartSLOduration=2.26589719 podStartE2EDuration="2.684200476s" podCreationTimestamp="2025-09-30 08:34:07 +0000 UTC" firstStartedPulling="2025-09-30 08:34:08.616727861 +0000 UTC m=+1872.068927138" lastFinishedPulling="2025-09-30 08:34:09.035031157 +0000 UTC m=+1872.487230424" observedRunningTime="2025-09-30 08:34:09.680350225 +0000 UTC m=+1873.132549502" watchObservedRunningTime="2025-09-30 08:34:09.684200476 +0000 UTC m=+1873.136399743" Sep 30 08:34:11 crc kubenswrapper[4810]: I0930 08:34:11.307626 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:34:11 crc kubenswrapper[4810]: E0930 08:34:11.308350 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:34:15 crc kubenswrapper[4810]: I0930 08:34:15.073992 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mhwt7"] Sep 30 08:34:15 crc kubenswrapper[4810]: I0930 08:34:15.092073 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mhwt7"] Sep 30 08:34:15 crc kubenswrapper[4810]: I0930 08:34:15.318827 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cdc8c6a-9e0f-47b0-888d-af03c4339799" path="/var/lib/kubelet/pods/6cdc8c6a-9e0f-47b0-888d-af03c4339799/volumes" Sep 30 08:34:20 crc kubenswrapper[4810]: I0930 08:34:20.324933 4810 scope.go:117] "RemoveContainer" containerID="17d1f548f5ae13f7f2fb2e80bcc3962ca7cf9e3bb3f204dc17076f9db1acdf15" Sep 30 08:34:20 crc kubenswrapper[4810]: I0930 08:34:20.363745 4810 scope.go:117] "RemoveContainer" containerID="0e01270242ec97d2e718beaa40e6f09952eddc7a17f4e3492bf1c883b2d0d5ec" Sep 30 08:34:20 crc kubenswrapper[4810]: I0930 08:34:20.450181 4810 scope.go:117] "RemoveContainer" containerID="73e5b89bd1e0fb17af98a0abd0af887cac1e4f3253865d48831c7d567fcaa6c5" Sep 30 08:34:20 crc kubenswrapper[4810]: I0930 08:34:20.480405 4810 scope.go:117] "RemoveContainer" containerID="8a4ccfcefec56c090ccac05acacdc1358a2cf3204037f22800b3da0ee707741d" Sep 30 08:34:20 crc kubenswrapper[4810]: I0930 08:34:20.532829 4810 scope.go:117] "RemoveContainer" containerID="36de72ecd40625a782355017c702907934cf07fad7f9841c55c526dd9260b553" Sep 30 08:34:20 crc kubenswrapper[4810]: I0930 08:34:20.577232 4810 scope.go:117] "RemoveContainer" containerID="2dd21650479bf94a3811552d259d60b9c28a18aa4c63d3352ff8e3e1adef3397" Sep 30 08:34:20 crc kubenswrapper[4810]: I0930 08:34:20.630696 4810 scope.go:117] "RemoveContainer" containerID="19ed5ba96885f13d5d5fd5ea34dd10a1a27ae7457de53eff539bdf7679e0114d" Sep 30 08:34:22 crc kubenswrapper[4810]: I0930 08:34:22.307437 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:34:22 crc kubenswrapper[4810]: E0930 08:34:22.308013 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:34:34 crc kubenswrapper[4810]: I0930 08:34:34.061192 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-xtgps"] Sep 30 08:34:34 crc kubenswrapper[4810]: I0930 08:34:34.091256 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-xtgps"] Sep 30 08:34:34 crc kubenswrapper[4810]: I0930 08:34:34.306997 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:34:34 crc kubenswrapper[4810]: E0930 08:34:34.307339 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:34:35 crc kubenswrapper[4810]: I0930 08:34:35.331098 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="376d03ec-c5fa-4f85-860f-2e6e659958ae" path="/var/lib/kubelet/pods/376d03ec-c5fa-4f85-860f-2e6e659958ae/volumes" Sep 30 08:34:39 crc kubenswrapper[4810]: I0930 08:34:39.039936 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7l4sp"] Sep 30 08:34:39 crc kubenswrapper[4810]: I0930 08:34:39.053015 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7l4sp"] Sep 30 08:34:39 crc kubenswrapper[4810]: I0930 08:34:39.320376 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="528064bb-fbc6-42fb-b46d-e932ad6971b1" path="/var/lib/kubelet/pods/528064bb-fbc6-42fb-b46d-e932ad6971b1/volumes" Sep 30 08:34:49 crc kubenswrapper[4810]: I0930 08:34:49.307551 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:34:49 crc kubenswrapper[4810]: E0930 08:34:49.308770 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:34:54 crc kubenswrapper[4810]: I0930 08:34:54.194778 4810 generic.go:334] "Generic (PLEG): container finished" podID="75b17470-760f-4fa6-9550-c2281ed5ae6b" containerID="fbba86a400be745516633e58871e85f3975e6f145b24825d72c60b684d992017" exitCode=0 Sep 30 08:34:54 crc kubenswrapper[4810]: I0930 08:34:54.194857 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" event={"ID":"75b17470-760f-4fa6-9550-c2281ed5ae6b","Type":"ContainerDied","Data":"fbba86a400be745516633e58871e85f3975e6f145b24825d72c60b684d992017"} Sep 30 08:34:55 crc kubenswrapper[4810]: I0930 08:34:55.627001 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:55 crc kubenswrapper[4810]: I0930 08:34:55.756178 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwkch\" (UniqueName: \"kubernetes.io/projected/75b17470-760f-4fa6-9550-c2281ed5ae6b-kube-api-access-pwkch\") pod \"75b17470-760f-4fa6-9550-c2281ed5ae6b\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " Sep 30 08:34:55 crc kubenswrapper[4810]: I0930 08:34:55.756297 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-inventory\") pod \"75b17470-760f-4fa6-9550-c2281ed5ae6b\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " Sep 30 08:34:55 crc kubenswrapper[4810]: I0930 08:34:55.756388 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-ssh-key\") pod \"75b17470-760f-4fa6-9550-c2281ed5ae6b\" (UID: \"75b17470-760f-4fa6-9550-c2281ed5ae6b\") " Sep 30 08:34:55 crc kubenswrapper[4810]: I0930 08:34:55.762806 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75b17470-760f-4fa6-9550-c2281ed5ae6b-kube-api-access-pwkch" (OuterVolumeSpecName: "kube-api-access-pwkch") pod "75b17470-760f-4fa6-9550-c2281ed5ae6b" (UID: "75b17470-760f-4fa6-9550-c2281ed5ae6b"). InnerVolumeSpecName "kube-api-access-pwkch". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:34:55 crc kubenswrapper[4810]: I0930 08:34:55.786412 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "75b17470-760f-4fa6-9550-c2281ed5ae6b" (UID: "75b17470-760f-4fa6-9550-c2281ed5ae6b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:34:55 crc kubenswrapper[4810]: I0930 08:34:55.811662 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-inventory" (OuterVolumeSpecName: "inventory") pod "75b17470-760f-4fa6-9550-c2281ed5ae6b" (UID: "75b17470-760f-4fa6-9550-c2281ed5ae6b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:34:55 crc kubenswrapper[4810]: I0930 08:34:55.860409 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:34:55 crc kubenswrapper[4810]: I0930 08:34:55.860730 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75b17470-760f-4fa6-9550-c2281ed5ae6b-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:34:55 crc kubenswrapper[4810]: I0930 08:34:55.860912 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwkch\" (UniqueName: \"kubernetes.io/projected/75b17470-760f-4fa6-9550-c2281ed5ae6b-kube-api-access-pwkch\") on node \"crc\" DevicePath \"\"" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.222583 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" event={"ID":"75b17470-760f-4fa6-9550-c2281ed5ae6b","Type":"ContainerDied","Data":"40eeb9b78dbc3bff2babf28609ba8f31deaff7030018862b0f5d4d026cbf11c8"} Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.222948 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40eeb9b78dbc3bff2babf28609ba8f31deaff7030018862b0f5d4d026cbf11c8" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.222667 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9gnn7" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.369633 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx"] Sep 30 08:34:56 crc kubenswrapper[4810]: E0930 08:34:56.370355 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75b17470-760f-4fa6-9550-c2281ed5ae6b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.370438 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="75b17470-760f-4fa6-9550-c2281ed5ae6b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.370684 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="75b17470-760f-4fa6-9550-c2281ed5ae6b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.371545 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.374155 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.378113 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.378321 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.378547 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.383761 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx"] Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.475220 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzncw\" (UniqueName: \"kubernetes.io/projected/80653750-cd3d-4e95-b4a1-4909c325c34c-kube-api-access-mzncw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.475376 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.475513 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.577463 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzncw\" (UniqueName: \"kubernetes.io/projected/80653750-cd3d-4e95-b4a1-4909c325c34c-kube-api-access-mzncw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.577729 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.577853 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.584817 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.596448 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.598015 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzncw\" (UniqueName: \"kubernetes.io/projected/80653750-cd3d-4e95-b4a1-4909c325c34c-kube-api-access-mzncw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:34:56 crc kubenswrapper[4810]: I0930 08:34:56.707241 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:34:57 crc kubenswrapper[4810]: I0930 08:34:57.345730 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx"] Sep 30 08:34:57 crc kubenswrapper[4810]: I0930 08:34:57.804401 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:34:58 crc kubenswrapper[4810]: I0930 08:34:58.247208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" event={"ID":"80653750-cd3d-4e95-b4a1-4909c325c34c","Type":"ContainerStarted","Data":"5730b32edc98847851ca31c6903f628f9018903a7a237e7ab3ff2066ddad6d5c"} Sep 30 08:34:58 crc kubenswrapper[4810]: I0930 08:34:58.247627 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" event={"ID":"80653750-cd3d-4e95-b4a1-4909c325c34c","Type":"ContainerStarted","Data":"3b2cab4ff7551def296d5c82a3186b5e16684e0bb5315ecf0d71a5e47aa95bda"} Sep 30 08:34:58 crc kubenswrapper[4810]: I0930 08:34:58.265144 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" podStartSLOduration=1.798336613 podStartE2EDuration="2.265126694s" podCreationTimestamp="2025-09-30 08:34:56 +0000 UTC" firstStartedPulling="2025-09-30 08:34:57.334972862 +0000 UTC m=+1920.787172159" lastFinishedPulling="2025-09-30 08:34:57.801762973 +0000 UTC m=+1921.253962240" observedRunningTime="2025-09-30 08:34:58.263471176 +0000 UTC m=+1921.715670443" watchObservedRunningTime="2025-09-30 08:34:58.265126694 +0000 UTC m=+1921.717325961" Sep 30 08:35:02 crc kubenswrapper[4810]: I0930 08:35:02.306664 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:35:02 crc kubenswrapper[4810]: E0930 08:35:02.307613 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:35:13 crc kubenswrapper[4810]: I0930 08:35:13.307210 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:35:13 crc kubenswrapper[4810]: E0930 08:35:13.307881 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:35:18 crc kubenswrapper[4810]: I0930 08:35:18.050054 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-pspj8"] Sep 30 08:35:18 crc kubenswrapper[4810]: I0930 08:35:18.062205 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-pspj8"] Sep 30 08:35:19 crc kubenswrapper[4810]: I0930 08:35:19.319914 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5df8e23-2303-4108-a55c-775633e3b4c9" path="/var/lib/kubelet/pods/a5df8e23-2303-4108-a55c-775633e3b4c9/volumes" Sep 30 08:35:20 crc kubenswrapper[4810]: I0930 08:35:20.849969 4810 scope.go:117] "RemoveContainer" containerID="26150a9bcc58d9e492f465ced6fab9fc13397d1412a95f47629adc56a5c55140" Sep 30 08:35:20 crc kubenswrapper[4810]: I0930 08:35:20.902292 4810 scope.go:117] "RemoveContainer" containerID="2068763a38313dfabb6c73703f63fd6736a6e3acea867c3941ae6cd2ee63af16" Sep 30 08:35:20 crc kubenswrapper[4810]: I0930 08:35:20.976185 4810 scope.go:117] "RemoveContainer" containerID="684c407e93bfe61796c8e02b7f5d80aaffe4c10d01d333e151b35197a331061b" Sep 30 08:35:25 crc kubenswrapper[4810]: I0930 08:35:25.306171 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:35:25 crc kubenswrapper[4810]: E0930 08:35:25.306849 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:35:40 crc kubenswrapper[4810]: I0930 08:35:40.308221 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:35:40 crc kubenswrapper[4810]: E0930 08:35:40.309765 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:35:52 crc kubenswrapper[4810]: I0930 08:35:52.306169 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:35:52 crc kubenswrapper[4810]: E0930 08:35:52.307091 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:35:59 crc kubenswrapper[4810]: I0930 08:35:59.049376 4810 generic.go:334] "Generic (PLEG): container finished" podID="80653750-cd3d-4e95-b4a1-4909c325c34c" containerID="5730b32edc98847851ca31c6903f628f9018903a7a237e7ab3ff2066ddad6d5c" exitCode=0 Sep 30 08:35:59 crc kubenswrapper[4810]: I0930 08:35:59.049476 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" event={"ID":"80653750-cd3d-4e95-b4a1-4909c325c34c","Type":"ContainerDied","Data":"5730b32edc98847851ca31c6903f628f9018903a7a237e7ab3ff2066ddad6d5c"} Sep 30 08:36:00 crc kubenswrapper[4810]: I0930 08:36:00.672349 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:36:00 crc kubenswrapper[4810]: I0930 08:36:00.810077 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-inventory\") pod \"80653750-cd3d-4e95-b4a1-4909c325c34c\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " Sep 30 08:36:00 crc kubenswrapper[4810]: I0930 08:36:00.810230 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-ssh-key\") pod \"80653750-cd3d-4e95-b4a1-4909c325c34c\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " Sep 30 08:36:00 crc kubenswrapper[4810]: I0930 08:36:00.810361 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzncw\" (UniqueName: \"kubernetes.io/projected/80653750-cd3d-4e95-b4a1-4909c325c34c-kube-api-access-mzncw\") pod \"80653750-cd3d-4e95-b4a1-4909c325c34c\" (UID: \"80653750-cd3d-4e95-b4a1-4909c325c34c\") " Sep 30 08:36:00 crc kubenswrapper[4810]: I0930 08:36:00.815967 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80653750-cd3d-4e95-b4a1-4909c325c34c-kube-api-access-mzncw" (OuterVolumeSpecName: "kube-api-access-mzncw") pod "80653750-cd3d-4e95-b4a1-4909c325c34c" (UID: "80653750-cd3d-4e95-b4a1-4909c325c34c"). InnerVolumeSpecName "kube-api-access-mzncw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:36:00 crc kubenswrapper[4810]: I0930 08:36:00.868601 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "80653750-cd3d-4e95-b4a1-4909c325c34c" (UID: "80653750-cd3d-4e95-b4a1-4909c325c34c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:36:00 crc kubenswrapper[4810]: I0930 08:36:00.869793 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-inventory" (OuterVolumeSpecName: "inventory") pod "80653750-cd3d-4e95-b4a1-4909c325c34c" (UID: "80653750-cd3d-4e95-b4a1-4909c325c34c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:36:00 crc kubenswrapper[4810]: I0930 08:36:00.913720 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:00 crc kubenswrapper[4810]: I0930 08:36:00.913766 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzncw\" (UniqueName: \"kubernetes.io/projected/80653750-cd3d-4e95-b4a1-4909c325c34c-kube-api-access-mzncw\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:00 crc kubenswrapper[4810]: I0930 08:36:00.913817 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80653750-cd3d-4e95-b4a1-4909c325c34c-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.073789 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" event={"ID":"80653750-cd3d-4e95-b4a1-4909c325c34c","Type":"ContainerDied","Data":"3b2cab4ff7551def296d5c82a3186b5e16684e0bb5315ecf0d71a5e47aa95bda"} Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.073832 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b2cab4ff7551def296d5c82a3186b5e16684e0bb5315ecf0d71a5e47aa95bda" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.073908 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.196702 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6gw9x"] Sep 30 08:36:01 crc kubenswrapper[4810]: E0930 08:36:01.197116 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80653750-cd3d-4e95-b4a1-4909c325c34c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.197134 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="80653750-cd3d-4e95-b4a1-4909c325c34c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.197344 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="80653750-cd3d-4e95-b4a1-4909c325c34c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.197978 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.200442 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.201339 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.201890 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.201924 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.221703 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6gw9x"] Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.322867 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6gw9x\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.322918 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6gw9x\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.322963 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78xxp\" (UniqueName: \"kubernetes.io/projected/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-kube-api-access-78xxp\") pod \"ssh-known-hosts-edpm-deployment-6gw9x\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.425107 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6gw9x\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.425173 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6gw9x\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.425228 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78xxp\" (UniqueName: \"kubernetes.io/projected/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-kube-api-access-78xxp\") pod \"ssh-known-hosts-edpm-deployment-6gw9x\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.430117 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6gw9x\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.438545 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6gw9x\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.447876 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78xxp\" (UniqueName: \"kubernetes.io/projected/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-kube-api-access-78xxp\") pod \"ssh-known-hosts-edpm-deployment-6gw9x\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:01 crc kubenswrapper[4810]: I0930 08:36:01.536712 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:02 crc kubenswrapper[4810]: I0930 08:36:02.139664 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6gw9x"] Sep 30 08:36:03 crc kubenswrapper[4810]: I0930 08:36:03.114087 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" event={"ID":"d5b6b2ff-6366-48ba-9328-7d99a9da7e08","Type":"ContainerStarted","Data":"f6bf089efe6e03072d661acabe23efc9f12009903a9531aa63f81a58053b8bd0"} Sep 30 08:36:03 crc kubenswrapper[4810]: I0930 08:36:03.306673 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:36:03 crc kubenswrapper[4810]: E0930 08:36:03.306998 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:36:04 crc kubenswrapper[4810]: I0930 08:36:04.132611 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" event={"ID":"d5b6b2ff-6366-48ba-9328-7d99a9da7e08","Type":"ContainerStarted","Data":"ba0e5a917ab6d39d06aacbec656bd0f0b605d31effcd118577a7e59d8d69fd96"} Sep 30 08:36:04 crc kubenswrapper[4810]: I0930 08:36:04.169948 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" podStartSLOduration=2.209745698 podStartE2EDuration="3.169922214s" podCreationTimestamp="2025-09-30 08:36:01 +0000 UTC" firstStartedPulling="2025-09-30 08:36:02.148536381 +0000 UTC m=+1985.600735648" lastFinishedPulling="2025-09-30 08:36:03.108712877 +0000 UTC m=+1986.560912164" observedRunningTime="2025-09-30 08:36:04.153816401 +0000 UTC m=+1987.606015698" watchObservedRunningTime="2025-09-30 08:36:04.169922214 +0000 UTC m=+1987.622121521" Sep 30 08:36:11 crc kubenswrapper[4810]: I0930 08:36:11.220613 4810 generic.go:334] "Generic (PLEG): container finished" podID="d5b6b2ff-6366-48ba-9328-7d99a9da7e08" containerID="ba0e5a917ab6d39d06aacbec656bd0f0b605d31effcd118577a7e59d8d69fd96" exitCode=0 Sep 30 08:36:11 crc kubenswrapper[4810]: I0930 08:36:11.220750 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" event={"ID":"d5b6b2ff-6366-48ba-9328-7d99a9da7e08","Type":"ContainerDied","Data":"ba0e5a917ab6d39d06aacbec656bd0f0b605d31effcd118577a7e59d8d69fd96"} Sep 30 08:36:12 crc kubenswrapper[4810]: I0930 08:36:12.693796 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:12 crc kubenswrapper[4810]: I0930 08:36:12.776504 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-inventory-0\") pod \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " Sep 30 08:36:12 crc kubenswrapper[4810]: I0930 08:36:12.776634 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78xxp\" (UniqueName: \"kubernetes.io/projected/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-kube-api-access-78xxp\") pod \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " Sep 30 08:36:12 crc kubenswrapper[4810]: I0930 08:36:12.785554 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-kube-api-access-78xxp" (OuterVolumeSpecName: "kube-api-access-78xxp") pod "d5b6b2ff-6366-48ba-9328-7d99a9da7e08" (UID: "d5b6b2ff-6366-48ba-9328-7d99a9da7e08"). InnerVolumeSpecName "kube-api-access-78xxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:36:12 crc kubenswrapper[4810]: I0930 08:36:12.847890 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "d5b6b2ff-6366-48ba-9328-7d99a9da7e08" (UID: "d5b6b2ff-6366-48ba-9328-7d99a9da7e08"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:36:12 crc kubenswrapper[4810]: I0930 08:36:12.878373 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-ssh-key-openstack-edpm-ipam\") pod \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\" (UID: \"d5b6b2ff-6366-48ba-9328-7d99a9da7e08\") " Sep 30 08:36:12 crc kubenswrapper[4810]: I0930 08:36:12.879903 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78xxp\" (UniqueName: \"kubernetes.io/projected/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-kube-api-access-78xxp\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:12 crc kubenswrapper[4810]: I0930 08:36:12.879962 4810 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-inventory-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:12 crc kubenswrapper[4810]: I0930 08:36:12.906438 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d5b6b2ff-6366-48ba-9328-7d99a9da7e08" (UID: "d5b6b2ff-6366-48ba-9328-7d99a9da7e08"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:36:12 crc kubenswrapper[4810]: I0930 08:36:12.981975 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d5b6b2ff-6366-48ba-9328-7d99a9da7e08-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.263740 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" event={"ID":"d5b6b2ff-6366-48ba-9328-7d99a9da7e08","Type":"ContainerDied","Data":"f6bf089efe6e03072d661acabe23efc9f12009903a9531aa63f81a58053b8bd0"} Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.263785 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6bf089efe6e03072d661acabe23efc9f12009903a9531aa63f81a58053b8bd0" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.263811 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6gw9x" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.351049 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn"] Sep 30 08:36:13 crc kubenswrapper[4810]: E0930 08:36:13.351834 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5b6b2ff-6366-48ba-9328-7d99a9da7e08" containerName="ssh-known-hosts-edpm-deployment" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.351873 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5b6b2ff-6366-48ba-9328-7d99a9da7e08" containerName="ssh-known-hosts-edpm-deployment" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.352296 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5b6b2ff-6366-48ba-9328-7d99a9da7e08" containerName="ssh-known-hosts-edpm-deployment" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.353538 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.356675 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.357627 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.359325 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.359337 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.368949 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn"] Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.391642 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v8zdn\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.391971 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v8zdn\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.392151 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz9rt\" (UniqueName: \"kubernetes.io/projected/48233a79-011d-4b63-8f0c-885c651ddbee-kube-api-access-zz9rt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v8zdn\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.495666 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v8zdn\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.495822 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v8zdn\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.495902 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz9rt\" (UniqueName: \"kubernetes.io/projected/48233a79-011d-4b63-8f0c-885c651ddbee-kube-api-access-zz9rt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v8zdn\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.500987 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v8zdn\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.502527 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v8zdn\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.517997 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz9rt\" (UniqueName: \"kubernetes.io/projected/48233a79-011d-4b63-8f0c-885c651ddbee-kube-api-access-zz9rt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-v8zdn\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:13 crc kubenswrapper[4810]: I0930 08:36:13.690362 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:14 crc kubenswrapper[4810]: I0930 08:36:14.332898 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn"] Sep 30 08:36:15 crc kubenswrapper[4810]: I0930 08:36:15.288253 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" event={"ID":"48233a79-011d-4b63-8f0c-885c651ddbee","Type":"ContainerStarted","Data":"a9f0a3cb27eab9457500a5166dbf9b64ec22f0eca56c8d8b1bfcda3fab5a83be"} Sep 30 08:36:15 crc kubenswrapper[4810]: I0930 08:36:15.289153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" event={"ID":"48233a79-011d-4b63-8f0c-885c651ddbee","Type":"ContainerStarted","Data":"51d582d46d5e160a0ad2334ff638aeec37d8a0470f2f2fe18652fac98816b15d"} Sep 30 08:36:15 crc kubenswrapper[4810]: I0930 08:36:15.308883 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" podStartSLOduration=1.612240185 podStartE2EDuration="2.308855038s" podCreationTimestamp="2025-09-30 08:36:13 +0000 UTC" firstStartedPulling="2025-09-30 08:36:14.331378593 +0000 UTC m=+1997.783577860" lastFinishedPulling="2025-09-30 08:36:15.027993446 +0000 UTC m=+1998.480192713" observedRunningTime="2025-09-30 08:36:15.307445516 +0000 UTC m=+1998.759644833" watchObservedRunningTime="2025-09-30 08:36:15.308855038 +0000 UTC m=+1998.761054335" Sep 30 08:36:18 crc kubenswrapper[4810]: I0930 08:36:18.307194 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:36:19 crc kubenswrapper[4810]: I0930 08:36:19.337100 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"04dd0426284d3a6d1a85a59da9e842cfb65296cf5e82dcf8080e9857a5eb8760"} Sep 30 08:36:26 crc kubenswrapper[4810]: I0930 08:36:26.421742 4810 generic.go:334] "Generic (PLEG): container finished" podID="48233a79-011d-4b63-8f0c-885c651ddbee" containerID="a9f0a3cb27eab9457500a5166dbf9b64ec22f0eca56c8d8b1bfcda3fab5a83be" exitCode=0 Sep 30 08:36:26 crc kubenswrapper[4810]: I0930 08:36:26.421850 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" event={"ID":"48233a79-011d-4b63-8f0c-885c651ddbee","Type":"ContainerDied","Data":"a9f0a3cb27eab9457500a5166dbf9b64ec22f0eca56c8d8b1bfcda3fab5a83be"} Sep 30 08:36:27 crc kubenswrapper[4810]: I0930 08:36:27.920247 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.015682 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-ssh-key\") pod \"48233a79-011d-4b63-8f0c-885c651ddbee\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.015726 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-inventory\") pod \"48233a79-011d-4b63-8f0c-885c651ddbee\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.015800 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz9rt\" (UniqueName: \"kubernetes.io/projected/48233a79-011d-4b63-8f0c-885c651ddbee-kube-api-access-zz9rt\") pod \"48233a79-011d-4b63-8f0c-885c651ddbee\" (UID: \"48233a79-011d-4b63-8f0c-885c651ddbee\") " Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.022997 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48233a79-011d-4b63-8f0c-885c651ddbee-kube-api-access-zz9rt" (OuterVolumeSpecName: "kube-api-access-zz9rt") pod "48233a79-011d-4b63-8f0c-885c651ddbee" (UID: "48233a79-011d-4b63-8f0c-885c651ddbee"). InnerVolumeSpecName "kube-api-access-zz9rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.051581 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "48233a79-011d-4b63-8f0c-885c651ddbee" (UID: "48233a79-011d-4b63-8f0c-885c651ddbee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.073596 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-inventory" (OuterVolumeSpecName: "inventory") pod "48233a79-011d-4b63-8f0c-885c651ddbee" (UID: "48233a79-011d-4b63-8f0c-885c651ddbee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.118690 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.118724 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48233a79-011d-4b63-8f0c-885c651ddbee-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.118736 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz9rt\" (UniqueName: \"kubernetes.io/projected/48233a79-011d-4b63-8f0c-885c651ddbee-kube-api-access-zz9rt\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.461013 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" event={"ID":"48233a79-011d-4b63-8f0c-885c651ddbee","Type":"ContainerDied","Data":"51d582d46d5e160a0ad2334ff638aeec37d8a0470f2f2fe18652fac98816b15d"} Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.461406 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51d582d46d5e160a0ad2334ff638aeec37d8a0470f2f2fe18652fac98816b15d" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.461120 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-v8zdn" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.533022 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp"] Sep 30 08:36:28 crc kubenswrapper[4810]: E0930 08:36:28.533516 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48233a79-011d-4b63-8f0c-885c651ddbee" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.533537 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="48233a79-011d-4b63-8f0c-885c651ddbee" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.533788 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="48233a79-011d-4b63-8f0c-885c651ddbee" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.534459 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.536589 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.540640 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.541247 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.541881 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.581164 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp"] Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.628061 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.628132 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpc88\" (UniqueName: \"kubernetes.io/projected/6da1c758-a003-49dd-95de-f47d13e11c0a-kube-api-access-hpc88\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.628199 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.729448 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.729527 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpc88\" (UniqueName: \"kubernetes.io/projected/6da1c758-a003-49dd-95de-f47d13e11c0a-kube-api-access-hpc88\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.729589 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.737208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.744478 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.748807 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpc88\" (UniqueName: \"kubernetes.io/projected/6da1c758-a003-49dd-95de-f47d13e11c0a-kube-api-access-hpc88\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:28 crc kubenswrapper[4810]: I0930 08:36:28.863967 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:29 crc kubenswrapper[4810]: I0930 08:36:29.496026 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp"] Sep 30 08:36:29 crc kubenswrapper[4810]: W0930 08:36:29.504732 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6da1c758_a003_49dd_95de_f47d13e11c0a.slice/crio-ada47e1c408f1e3f9cc0bf884c15d01ffc51dd6e95b0ae27ed372ee909b012b9 WatchSource:0}: Error finding container ada47e1c408f1e3f9cc0bf884c15d01ffc51dd6e95b0ae27ed372ee909b012b9: Status 404 returned error can't find the container with id ada47e1c408f1e3f9cc0bf884c15d01ffc51dd6e95b0ae27ed372ee909b012b9 Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.491982 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7jq4v"] Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.495732 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.502094 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" event={"ID":"6da1c758-a003-49dd-95de-f47d13e11c0a","Type":"ContainerStarted","Data":"002b952cca0e194163d343cc3004ead40da7ecc0a99ccab288603248a89d575b"} Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.502155 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" event={"ID":"6da1c758-a003-49dd-95de-f47d13e11c0a","Type":"ContainerStarted","Data":"ada47e1c408f1e3f9cc0bf884c15d01ffc51dd6e95b0ae27ed372ee909b012b9"} Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.511714 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7jq4v"] Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.562620 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" podStartSLOduration=1.9073024840000001 podStartE2EDuration="2.562591458s" podCreationTimestamp="2025-09-30 08:36:28 +0000 UTC" firstStartedPulling="2025-09-30 08:36:29.508142488 +0000 UTC m=+2012.960341755" lastFinishedPulling="2025-09-30 08:36:30.163431452 +0000 UTC m=+2013.615630729" observedRunningTime="2025-09-30 08:36:30.546245821 +0000 UTC m=+2013.998445138" watchObservedRunningTime="2025-09-30 08:36:30.562591458 +0000 UTC m=+2014.014790765" Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.593176 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-catalog-content\") pod \"certified-operators-7jq4v\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.593244 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-utilities\") pod \"certified-operators-7jq4v\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.593423 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr9jb\" (UniqueName: \"kubernetes.io/projected/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-kube-api-access-vr9jb\") pod \"certified-operators-7jq4v\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.694811 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr9jb\" (UniqueName: \"kubernetes.io/projected/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-kube-api-access-vr9jb\") pod \"certified-operators-7jq4v\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.694922 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-catalog-content\") pod \"certified-operators-7jq4v\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.694965 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-utilities\") pod \"certified-operators-7jq4v\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.695517 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-catalog-content\") pod \"certified-operators-7jq4v\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.695594 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-utilities\") pod \"certified-operators-7jq4v\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.735115 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr9jb\" (UniqueName: \"kubernetes.io/projected/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-kube-api-access-vr9jb\") pod \"certified-operators-7jq4v\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:30 crc kubenswrapper[4810]: I0930 08:36:30.824549 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:31 crc kubenswrapper[4810]: W0930 08:36:31.310952 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e9ae8b8_41fe_46e0_ba18_930fb7fa94a8.slice/crio-0d71cb321e0864dbb1ee139532a2c02889061e65e4e0126a0e294ced3ea254c8 WatchSource:0}: Error finding container 0d71cb321e0864dbb1ee139532a2c02889061e65e4e0126a0e294ced3ea254c8: Status 404 returned error can't find the container with id 0d71cb321e0864dbb1ee139532a2c02889061e65e4e0126a0e294ced3ea254c8 Sep 30 08:36:31 crc kubenswrapper[4810]: I0930 08:36:31.320601 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7jq4v"] Sep 30 08:36:31 crc kubenswrapper[4810]: I0930 08:36:31.514388 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jq4v" event={"ID":"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8","Type":"ContainerStarted","Data":"eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5"} Sep 30 08:36:31 crc kubenswrapper[4810]: I0930 08:36:31.514426 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jq4v" event={"ID":"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8","Type":"ContainerStarted","Data":"0d71cb321e0864dbb1ee139532a2c02889061e65e4e0126a0e294ced3ea254c8"} Sep 30 08:36:32 crc kubenswrapper[4810]: I0930 08:36:32.528305 4810 generic.go:334] "Generic (PLEG): container finished" podID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" containerID="eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5" exitCode=0 Sep 30 08:36:32 crc kubenswrapper[4810]: I0930 08:36:32.528382 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jq4v" event={"ID":"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8","Type":"ContainerDied","Data":"eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5"} Sep 30 08:36:34 crc kubenswrapper[4810]: I0930 08:36:34.553601 4810 generic.go:334] "Generic (PLEG): container finished" podID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" containerID="568e401fcce6221982d384bc80c71d032eb0b33d1a8268e4eaee757593892175" exitCode=0 Sep 30 08:36:34 crc kubenswrapper[4810]: I0930 08:36:34.553732 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jq4v" event={"ID":"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8","Type":"ContainerDied","Data":"568e401fcce6221982d384bc80c71d032eb0b33d1a8268e4eaee757593892175"} Sep 30 08:36:35 crc kubenswrapper[4810]: I0930 08:36:35.569708 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jq4v" event={"ID":"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8","Type":"ContainerStarted","Data":"7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2"} Sep 30 08:36:35 crc kubenswrapper[4810]: I0930 08:36:35.615127 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7jq4v" podStartSLOduration=3.144272668 podStartE2EDuration="5.615095766s" podCreationTimestamp="2025-09-30 08:36:30 +0000 UTC" firstStartedPulling="2025-09-30 08:36:32.531055727 +0000 UTC m=+2015.983254994" lastFinishedPulling="2025-09-30 08:36:35.001878815 +0000 UTC m=+2018.454078092" observedRunningTime="2025-09-30 08:36:35.59240472 +0000 UTC m=+2019.044603987" watchObservedRunningTime="2025-09-30 08:36:35.615095766 +0000 UTC m=+2019.067295073" Sep 30 08:36:40 crc kubenswrapper[4810]: I0930 08:36:40.824962 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:40 crc kubenswrapper[4810]: I0930 08:36:40.825642 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:40 crc kubenswrapper[4810]: I0930 08:36:40.899478 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:41 crc kubenswrapper[4810]: I0930 08:36:41.680699 4810 generic.go:334] "Generic (PLEG): container finished" podID="6da1c758-a003-49dd-95de-f47d13e11c0a" containerID="002b952cca0e194163d343cc3004ead40da7ecc0a99ccab288603248a89d575b" exitCode=0 Sep 30 08:36:41 crc kubenswrapper[4810]: I0930 08:36:41.680816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" event={"ID":"6da1c758-a003-49dd-95de-f47d13e11c0a","Type":"ContainerDied","Data":"002b952cca0e194163d343cc3004ead40da7ecc0a99ccab288603248a89d575b"} Sep 30 08:36:41 crc kubenswrapper[4810]: I0930 08:36:41.744902 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:41 crc kubenswrapper[4810]: I0930 08:36:41.809090 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7jq4v"] Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.222865 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.406584 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-ssh-key\") pod \"6da1c758-a003-49dd-95de-f47d13e11c0a\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.406860 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpc88\" (UniqueName: \"kubernetes.io/projected/6da1c758-a003-49dd-95de-f47d13e11c0a-kube-api-access-hpc88\") pod \"6da1c758-a003-49dd-95de-f47d13e11c0a\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.406981 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-inventory\") pod \"6da1c758-a003-49dd-95de-f47d13e11c0a\" (UID: \"6da1c758-a003-49dd-95de-f47d13e11c0a\") " Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.414537 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da1c758-a003-49dd-95de-f47d13e11c0a-kube-api-access-hpc88" (OuterVolumeSpecName: "kube-api-access-hpc88") pod "6da1c758-a003-49dd-95de-f47d13e11c0a" (UID: "6da1c758-a003-49dd-95de-f47d13e11c0a"). InnerVolumeSpecName "kube-api-access-hpc88". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.435166 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6da1c758-a003-49dd-95de-f47d13e11c0a" (UID: "6da1c758-a003-49dd-95de-f47d13e11c0a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.453483 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-inventory" (OuterVolumeSpecName: "inventory") pod "6da1c758-a003-49dd-95de-f47d13e11c0a" (UID: "6da1c758-a003-49dd-95de-f47d13e11c0a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.511133 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.511185 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpc88\" (UniqueName: \"kubernetes.io/projected/6da1c758-a003-49dd-95de-f47d13e11c0a-kube-api-access-hpc88\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.511204 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6da1c758-a003-49dd-95de-f47d13e11c0a-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.715788 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7jq4v" podUID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" containerName="registry-server" containerID="cri-o://7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2" gracePeriod=2 Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.716253 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.716333 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp" event={"ID":"6da1c758-a003-49dd-95de-f47d13e11c0a","Type":"ContainerDied","Data":"ada47e1c408f1e3f9cc0bf884c15d01ffc51dd6e95b0ae27ed372ee909b012b9"} Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.716386 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ada47e1c408f1e3f9cc0bf884c15d01ffc51dd6e95b0ae27ed372ee909b012b9" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.858481 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf"] Sep 30 08:36:43 crc kubenswrapper[4810]: E0930 08:36:43.859389 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da1c758-a003-49dd-95de-f47d13e11c0a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.859408 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da1c758-a003-49dd-95de-f47d13e11c0a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.859690 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da1c758-a003-49dd-95de-f47d13e11c0a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.860681 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.865090 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.865424 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.865839 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.866139 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.866666 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.866863 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.867077 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.869292 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Sep 30 08:36:43 crc kubenswrapper[4810]: I0930 08:36:43.879164 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf"] Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022267 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022363 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022410 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022456 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022502 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022547 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022572 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022642 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022702 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022807 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022834 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022893 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022915 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg88m\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-kube-api-access-bg88m\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.022994 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.124565 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.124668 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125017 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125081 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125141 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125194 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125269 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125334 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125412 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125476 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125578 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125617 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125686 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.125719 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg88m\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-kube-api-access-bg88m\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.130492 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.132458 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.133489 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.134879 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.142230 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.143213 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.144132 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.145529 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.145899 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.146770 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.146883 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.162807 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.163176 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.163391 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg88m\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-kube-api-access-bg88m\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2rltf\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.246815 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.367904 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.534330 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-utilities\") pod \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.534560 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-catalog-content\") pod \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.534745 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr9jb\" (UniqueName: \"kubernetes.io/projected/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-kube-api-access-vr9jb\") pod \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\" (UID: \"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8\") " Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.536809 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-utilities" (OuterVolumeSpecName: "utilities") pod "3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" (UID: "3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.540491 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-kube-api-access-vr9jb" (OuterVolumeSpecName: "kube-api-access-vr9jb") pod "3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" (UID: "3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8"). InnerVolumeSpecName "kube-api-access-vr9jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.576409 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" (UID: "3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.637161 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr9jb\" (UniqueName: \"kubernetes.io/projected/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-kube-api-access-vr9jb\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.637196 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.637207 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.732190 4810 generic.go:334] "Generic (PLEG): container finished" podID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" containerID="7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2" exitCode=0 Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.732575 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jq4v" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.732461 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jq4v" event={"ID":"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8","Type":"ContainerDied","Data":"7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2"} Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.732750 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jq4v" event={"ID":"3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8","Type":"ContainerDied","Data":"0d71cb321e0864dbb1ee139532a2c02889061e65e4e0126a0e294ced3ea254c8"} Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.732775 4810 scope.go:117] "RemoveContainer" containerID="7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.777668 4810 scope.go:117] "RemoveContainer" containerID="568e401fcce6221982d384bc80c71d032eb0b33d1a8268e4eaee757593892175" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.807850 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7jq4v"] Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.819031 4810 scope.go:117] "RemoveContainer" containerID="eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.822173 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7jq4v"] Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.849053 4810 scope.go:117] "RemoveContainer" containerID="7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2" Sep 30 08:36:44 crc kubenswrapper[4810]: E0930 08:36:44.849736 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2\": container with ID starting with 7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2 not found: ID does not exist" containerID="7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.849886 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2"} err="failed to get container status \"7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2\": rpc error: code = NotFound desc = could not find container \"7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2\": container with ID starting with 7e4273960a4d42f9b309a8562edadc9caf5cab4257b4f0768a0d08558d3ef6c2 not found: ID does not exist" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.849934 4810 scope.go:117] "RemoveContainer" containerID="568e401fcce6221982d384bc80c71d032eb0b33d1a8268e4eaee757593892175" Sep 30 08:36:44 crc kubenswrapper[4810]: E0930 08:36:44.850673 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"568e401fcce6221982d384bc80c71d032eb0b33d1a8268e4eaee757593892175\": container with ID starting with 568e401fcce6221982d384bc80c71d032eb0b33d1a8268e4eaee757593892175 not found: ID does not exist" containerID="568e401fcce6221982d384bc80c71d032eb0b33d1a8268e4eaee757593892175" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.850732 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"568e401fcce6221982d384bc80c71d032eb0b33d1a8268e4eaee757593892175"} err="failed to get container status \"568e401fcce6221982d384bc80c71d032eb0b33d1a8268e4eaee757593892175\": rpc error: code = NotFound desc = could not find container \"568e401fcce6221982d384bc80c71d032eb0b33d1a8268e4eaee757593892175\": container with ID starting with 568e401fcce6221982d384bc80c71d032eb0b33d1a8268e4eaee757593892175 not found: ID does not exist" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.850781 4810 scope.go:117] "RemoveContainer" containerID="eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.850756 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf"] Sep 30 08:36:44 crc kubenswrapper[4810]: E0930 08:36:44.851241 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5\": container with ID starting with eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5 not found: ID does not exist" containerID="eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5" Sep 30 08:36:44 crc kubenswrapper[4810]: I0930 08:36:44.851333 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5"} err="failed to get container status \"eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5\": rpc error: code = NotFound desc = could not find container \"eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5\": container with ID starting with eb51da6befc35d0140f665f10ce0fe4f74641e41bbf51383d4f756f4bb3e4dd5 not found: ID does not exist" Sep 30 08:36:44 crc kubenswrapper[4810]: W0930 08:36:44.860622 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2e22e53_6826_4a58_abb0_eb2d279d2b1c.slice/crio-a24f250f208f11051162908847497fa896aa3c6ee98fd3689858cf121bcf6570 WatchSource:0}: Error finding container a24f250f208f11051162908847497fa896aa3c6ee98fd3689858cf121bcf6570: Status 404 returned error can't find the container with id a24f250f208f11051162908847497fa896aa3c6ee98fd3689858cf121bcf6570 Sep 30 08:36:45 crc kubenswrapper[4810]: I0930 08:36:45.326632 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" path="/var/lib/kubelet/pods/3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8/volumes" Sep 30 08:36:45 crc kubenswrapper[4810]: I0930 08:36:45.746225 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" event={"ID":"d2e22e53-6826-4a58-abb0-eb2d279d2b1c","Type":"ContainerStarted","Data":"a027699df9d176ea07182e5039beb7f45e7f4e68b525aa45554713ab640c5acf"} Sep 30 08:36:45 crc kubenswrapper[4810]: I0930 08:36:45.746572 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" event={"ID":"d2e22e53-6826-4a58-abb0-eb2d279d2b1c","Type":"ContainerStarted","Data":"a24f250f208f11051162908847497fa896aa3c6ee98fd3689858cf121bcf6570"} Sep 30 08:36:45 crc kubenswrapper[4810]: I0930 08:36:45.773712 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" podStartSLOduration=2.179614346 podStartE2EDuration="2.773689617s" podCreationTimestamp="2025-09-30 08:36:43 +0000 UTC" firstStartedPulling="2025-09-30 08:36:44.863833092 +0000 UTC m=+2028.316032359" lastFinishedPulling="2025-09-30 08:36:45.457908323 +0000 UTC m=+2028.910107630" observedRunningTime="2025-09-30 08:36:45.767252665 +0000 UTC m=+2029.219451932" watchObservedRunningTime="2025-09-30 08:36:45.773689617 +0000 UTC m=+2029.225888894" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.590467 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m4h2d"] Sep 30 08:37:04 crc kubenswrapper[4810]: E0930 08:37:04.591532 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" containerName="extract-content" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.591551 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" containerName="extract-content" Sep 30 08:37:04 crc kubenswrapper[4810]: E0930 08:37:04.591584 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" containerName="extract-utilities" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.591595 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" containerName="extract-utilities" Sep 30 08:37:04 crc kubenswrapper[4810]: E0930 08:37:04.591639 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" containerName="registry-server" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.591648 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" containerName="registry-server" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.591890 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e9ae8b8-41fe-46e0-ba18-930fb7fa94a8" containerName="registry-server" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.593765 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.607198 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m4h2d"] Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.703104 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-catalog-content\") pod \"community-operators-m4h2d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.703359 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bx65\" (UniqueName: \"kubernetes.io/projected/a773ed55-235a-48d2-834c-96447542c35d-kube-api-access-6bx65\") pod \"community-operators-m4h2d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.703446 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-utilities\") pod \"community-operators-m4h2d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.805651 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-catalog-content\") pod \"community-operators-m4h2d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.805763 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bx65\" (UniqueName: \"kubernetes.io/projected/a773ed55-235a-48d2-834c-96447542c35d-kube-api-access-6bx65\") pod \"community-operators-m4h2d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.805801 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-utilities\") pod \"community-operators-m4h2d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.806296 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-utilities\") pod \"community-operators-m4h2d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.806494 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-catalog-content\") pod \"community-operators-m4h2d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.836338 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bx65\" (UniqueName: \"kubernetes.io/projected/a773ed55-235a-48d2-834c-96447542c35d-kube-api-access-6bx65\") pod \"community-operators-m4h2d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:04 crc kubenswrapper[4810]: I0930 08:37:04.929826 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:05 crc kubenswrapper[4810]: I0930 08:37:05.479698 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m4h2d"] Sep 30 08:37:05 crc kubenswrapper[4810]: W0930 08:37:05.492800 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda773ed55_235a_48d2_834c_96447542c35d.slice/crio-c0a6c0576c8352f988002cf1613691ec2b4aa216d0b6036184d9267241475365 WatchSource:0}: Error finding container c0a6c0576c8352f988002cf1613691ec2b4aa216d0b6036184d9267241475365: Status 404 returned error can't find the container with id c0a6c0576c8352f988002cf1613691ec2b4aa216d0b6036184d9267241475365 Sep 30 08:37:05 crc kubenswrapper[4810]: I0930 08:37:05.988054 4810 generic.go:334] "Generic (PLEG): container finished" podID="a773ed55-235a-48d2-834c-96447542c35d" containerID="9752fee604b50a14e25db74a50623f60855f39b344b7e8e79431b470a0ffc206" exitCode=0 Sep 30 08:37:05 crc kubenswrapper[4810]: I0930 08:37:05.988172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4h2d" event={"ID":"a773ed55-235a-48d2-834c-96447542c35d","Type":"ContainerDied","Data":"9752fee604b50a14e25db74a50623f60855f39b344b7e8e79431b470a0ffc206"} Sep 30 08:37:05 crc kubenswrapper[4810]: I0930 08:37:05.988329 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4h2d" event={"ID":"a773ed55-235a-48d2-834c-96447542c35d","Type":"ContainerStarted","Data":"c0a6c0576c8352f988002cf1613691ec2b4aa216d0b6036184d9267241475365"} Sep 30 08:37:07 crc kubenswrapper[4810]: I0930 08:37:07.010245 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4h2d" event={"ID":"a773ed55-235a-48d2-834c-96447542c35d","Type":"ContainerStarted","Data":"f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f"} Sep 30 08:37:08 crc kubenswrapper[4810]: I0930 08:37:08.035661 4810 generic.go:334] "Generic (PLEG): container finished" podID="a773ed55-235a-48d2-834c-96447542c35d" containerID="f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f" exitCode=0 Sep 30 08:37:08 crc kubenswrapper[4810]: I0930 08:37:08.035796 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4h2d" event={"ID":"a773ed55-235a-48d2-834c-96447542c35d","Type":"ContainerDied","Data":"f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f"} Sep 30 08:37:09 crc kubenswrapper[4810]: I0930 08:37:09.051521 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4h2d" event={"ID":"a773ed55-235a-48d2-834c-96447542c35d","Type":"ContainerStarted","Data":"31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a"} Sep 30 08:37:09 crc kubenswrapper[4810]: I0930 08:37:09.077002 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m4h2d" podStartSLOduration=2.532187931 podStartE2EDuration="5.076978841s" podCreationTimestamp="2025-09-30 08:37:04 +0000 UTC" firstStartedPulling="2025-09-30 08:37:05.992061336 +0000 UTC m=+2049.444260613" lastFinishedPulling="2025-09-30 08:37:08.536852256 +0000 UTC m=+2051.989051523" observedRunningTime="2025-09-30 08:37:09.071648592 +0000 UTC m=+2052.523847869" watchObservedRunningTime="2025-09-30 08:37:09.076978841 +0000 UTC m=+2052.529178108" Sep 30 08:37:14 crc kubenswrapper[4810]: I0930 08:37:14.930910 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:14 crc kubenswrapper[4810]: I0930 08:37:14.931381 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:15 crc kubenswrapper[4810]: I0930 08:37:15.022733 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:15 crc kubenswrapper[4810]: I0930 08:37:15.154587 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:15 crc kubenswrapper[4810]: I0930 08:37:15.260353 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m4h2d"] Sep 30 08:37:17 crc kubenswrapper[4810]: I0930 08:37:17.130165 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m4h2d" podUID="a773ed55-235a-48d2-834c-96447542c35d" containerName="registry-server" containerID="cri-o://31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a" gracePeriod=2 Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.051096 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.107031 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bx65\" (UniqueName: \"kubernetes.io/projected/a773ed55-235a-48d2-834c-96447542c35d-kube-api-access-6bx65\") pod \"a773ed55-235a-48d2-834c-96447542c35d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.107113 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-catalog-content\") pod \"a773ed55-235a-48d2-834c-96447542c35d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.107320 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-utilities\") pod \"a773ed55-235a-48d2-834c-96447542c35d\" (UID: \"a773ed55-235a-48d2-834c-96447542c35d\") " Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.108512 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-utilities" (OuterVolumeSpecName: "utilities") pod "a773ed55-235a-48d2-834c-96447542c35d" (UID: "a773ed55-235a-48d2-834c-96447542c35d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.119923 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a773ed55-235a-48d2-834c-96447542c35d-kube-api-access-6bx65" (OuterVolumeSpecName: "kube-api-access-6bx65") pod "a773ed55-235a-48d2-834c-96447542c35d" (UID: "a773ed55-235a-48d2-834c-96447542c35d"). InnerVolumeSpecName "kube-api-access-6bx65". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.147448 4810 generic.go:334] "Generic (PLEG): container finished" podID="a773ed55-235a-48d2-834c-96447542c35d" containerID="31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a" exitCode=0 Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.147499 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4h2d" event={"ID":"a773ed55-235a-48d2-834c-96447542c35d","Type":"ContainerDied","Data":"31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a"} Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.147532 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4h2d" event={"ID":"a773ed55-235a-48d2-834c-96447542c35d","Type":"ContainerDied","Data":"c0a6c0576c8352f988002cf1613691ec2b4aa216d0b6036184d9267241475365"} Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.147540 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4h2d" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.147552 4810 scope.go:117] "RemoveContainer" containerID="31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.182662 4810 scope.go:117] "RemoveContainer" containerID="f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.183316 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a773ed55-235a-48d2-834c-96447542c35d" (UID: "a773ed55-235a-48d2-834c-96447542c35d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.207599 4810 scope.go:117] "RemoveContainer" containerID="9752fee604b50a14e25db74a50623f60855f39b344b7e8e79431b470a0ffc206" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.209807 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bx65\" (UniqueName: \"kubernetes.io/projected/a773ed55-235a-48d2-834c-96447542c35d-kube-api-access-6bx65\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.209845 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.209859 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a773ed55-235a-48d2-834c-96447542c35d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.249815 4810 scope.go:117] "RemoveContainer" containerID="31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a" Sep 30 08:37:18 crc kubenswrapper[4810]: E0930 08:37:18.250251 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a\": container with ID starting with 31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a not found: ID does not exist" containerID="31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.250329 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a"} err="failed to get container status \"31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a\": rpc error: code = NotFound desc = could not find container \"31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a\": container with ID starting with 31131d64838026f38e3fb7ac66f840371ee2979148b1309a839f3a8d69a4f01a not found: ID does not exist" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.250352 4810 scope.go:117] "RemoveContainer" containerID="f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f" Sep 30 08:37:18 crc kubenswrapper[4810]: E0930 08:37:18.250856 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f\": container with ID starting with f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f not found: ID does not exist" containerID="f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.250884 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f"} err="failed to get container status \"f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f\": rpc error: code = NotFound desc = could not find container \"f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f\": container with ID starting with f0a8044a7e47f0887a59caa8d8588b0fb7dc6182db9a2bc672056b834991970f not found: ID does not exist" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.250899 4810 scope.go:117] "RemoveContainer" containerID="9752fee604b50a14e25db74a50623f60855f39b344b7e8e79431b470a0ffc206" Sep 30 08:37:18 crc kubenswrapper[4810]: E0930 08:37:18.251362 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9752fee604b50a14e25db74a50623f60855f39b344b7e8e79431b470a0ffc206\": container with ID starting with 9752fee604b50a14e25db74a50623f60855f39b344b7e8e79431b470a0ffc206 not found: ID does not exist" containerID="9752fee604b50a14e25db74a50623f60855f39b344b7e8e79431b470a0ffc206" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.251420 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9752fee604b50a14e25db74a50623f60855f39b344b7e8e79431b470a0ffc206"} err="failed to get container status \"9752fee604b50a14e25db74a50623f60855f39b344b7e8e79431b470a0ffc206\": rpc error: code = NotFound desc = could not find container \"9752fee604b50a14e25db74a50623f60855f39b344b7e8e79431b470a0ffc206\": container with ID starting with 9752fee604b50a14e25db74a50623f60855f39b344b7e8e79431b470a0ffc206 not found: ID does not exist" Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.481537 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m4h2d"] Sep 30 08:37:18 crc kubenswrapper[4810]: I0930 08:37:18.488983 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m4h2d"] Sep 30 08:37:19 crc kubenswrapper[4810]: I0930 08:37:19.319929 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a773ed55-235a-48d2-834c-96447542c35d" path="/var/lib/kubelet/pods/a773ed55-235a-48d2-834c-96447542c35d/volumes" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.657347 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gd9rg"] Sep 30 08:37:22 crc kubenswrapper[4810]: E0930 08:37:22.658389 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a773ed55-235a-48d2-834c-96447542c35d" containerName="extract-content" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.658410 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a773ed55-235a-48d2-834c-96447542c35d" containerName="extract-content" Sep 30 08:37:22 crc kubenswrapper[4810]: E0930 08:37:22.658433 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a773ed55-235a-48d2-834c-96447542c35d" containerName="extract-utilities" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.658444 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a773ed55-235a-48d2-834c-96447542c35d" containerName="extract-utilities" Sep 30 08:37:22 crc kubenswrapper[4810]: E0930 08:37:22.658485 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a773ed55-235a-48d2-834c-96447542c35d" containerName="registry-server" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.658494 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a773ed55-235a-48d2-834c-96447542c35d" containerName="registry-server" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.658812 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a773ed55-235a-48d2-834c-96447542c35d" containerName="registry-server" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.660799 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.673010 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gd9rg"] Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.700176 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwvfs\" (UniqueName: \"kubernetes.io/projected/fa76491c-76b2-446b-b46b-489fb26c062e-kube-api-access-hwvfs\") pod \"redhat-operators-gd9rg\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.700235 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-catalog-content\") pod \"redhat-operators-gd9rg\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.700418 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-utilities\") pod \"redhat-operators-gd9rg\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.802327 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwvfs\" (UniqueName: \"kubernetes.io/projected/fa76491c-76b2-446b-b46b-489fb26c062e-kube-api-access-hwvfs\") pod \"redhat-operators-gd9rg\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.802974 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-catalog-content\") pod \"redhat-operators-gd9rg\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.803364 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-catalog-content\") pod \"redhat-operators-gd9rg\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.803417 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-utilities\") pod \"redhat-operators-gd9rg\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.803780 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-utilities\") pod \"redhat-operators-gd9rg\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.821561 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwvfs\" (UniqueName: \"kubernetes.io/projected/fa76491c-76b2-446b-b46b-489fb26c062e-kube-api-access-hwvfs\") pod \"redhat-operators-gd9rg\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:22 crc kubenswrapper[4810]: I0930 08:37:22.993745 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:23 crc kubenswrapper[4810]: W0930 08:37:23.472019 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa76491c_76b2_446b_b46b_489fb26c062e.slice/crio-2cbe6f46ae9c93b7a7da09599462c520259bf2cb3d6567161a4347d507697bc6 WatchSource:0}: Error finding container 2cbe6f46ae9c93b7a7da09599462c520259bf2cb3d6567161a4347d507697bc6: Status 404 returned error can't find the container with id 2cbe6f46ae9c93b7a7da09599462c520259bf2cb3d6567161a4347d507697bc6 Sep 30 08:37:23 crc kubenswrapper[4810]: I0930 08:37:23.472818 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gd9rg"] Sep 30 08:37:24 crc kubenswrapper[4810]: I0930 08:37:24.221674 4810 generic.go:334] "Generic (PLEG): container finished" podID="fa76491c-76b2-446b-b46b-489fb26c062e" containerID="747ffe1618e805045cb758cf85a57e90d9726c2ea8a46d6733108913bdcd932a" exitCode=0 Sep 30 08:37:24 crc kubenswrapper[4810]: I0930 08:37:24.221803 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd9rg" event={"ID":"fa76491c-76b2-446b-b46b-489fb26c062e","Type":"ContainerDied","Data":"747ffe1618e805045cb758cf85a57e90d9726c2ea8a46d6733108913bdcd932a"} Sep 30 08:37:24 crc kubenswrapper[4810]: I0930 08:37:24.221898 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd9rg" event={"ID":"fa76491c-76b2-446b-b46b-489fb26c062e","Type":"ContainerStarted","Data":"2cbe6f46ae9c93b7a7da09599462c520259bf2cb3d6567161a4347d507697bc6"} Sep 30 08:37:25 crc kubenswrapper[4810]: I0930 08:37:25.233585 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd9rg" event={"ID":"fa76491c-76b2-446b-b46b-489fb26c062e","Type":"ContainerStarted","Data":"ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7"} Sep 30 08:37:27 crc kubenswrapper[4810]: I0930 08:37:27.254173 4810 generic.go:334] "Generic (PLEG): container finished" podID="fa76491c-76b2-446b-b46b-489fb26c062e" containerID="ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7" exitCode=0 Sep 30 08:37:27 crc kubenswrapper[4810]: I0930 08:37:27.254258 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd9rg" event={"ID":"fa76491c-76b2-446b-b46b-489fb26c062e","Type":"ContainerDied","Data":"ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7"} Sep 30 08:37:29 crc kubenswrapper[4810]: I0930 08:37:29.281734 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd9rg" event={"ID":"fa76491c-76b2-446b-b46b-489fb26c062e","Type":"ContainerStarted","Data":"d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66"} Sep 30 08:37:29 crc kubenswrapper[4810]: I0930 08:37:29.306068 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gd9rg" podStartSLOduration=3.161296442 podStartE2EDuration="7.306043078s" podCreationTimestamp="2025-09-30 08:37:22 +0000 UTC" firstStartedPulling="2025-09-30 08:37:24.224174505 +0000 UTC m=+2067.676373802" lastFinishedPulling="2025-09-30 08:37:28.368921171 +0000 UTC m=+2071.821120438" observedRunningTime="2025-09-30 08:37:29.302511212 +0000 UTC m=+2072.754710479" watchObservedRunningTime="2025-09-30 08:37:29.306043078 +0000 UTC m=+2072.758242365" Sep 30 08:37:30 crc kubenswrapper[4810]: I0930 08:37:30.294163 4810 generic.go:334] "Generic (PLEG): container finished" podID="d2e22e53-6826-4a58-abb0-eb2d279d2b1c" containerID="a027699df9d176ea07182e5039beb7f45e7f4e68b525aa45554713ab640c5acf" exitCode=0 Sep 30 08:37:30 crc kubenswrapper[4810]: I0930 08:37:30.294245 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" event={"ID":"d2e22e53-6826-4a58-abb0-eb2d279d2b1c","Type":"ContainerDied","Data":"a027699df9d176ea07182e5039beb7f45e7f4e68b525aa45554713ab640c5acf"} Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.812091 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976447 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976529 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-neutron-metadata-combined-ca-bundle\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976612 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-repo-setup-combined-ca-bundle\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976637 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976658 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-bootstrap-combined-ca-bundle\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976681 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg88m\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-kube-api-access-bg88m\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976698 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ssh-key\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976715 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-inventory\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976745 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976765 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-nova-combined-ca-bundle\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976790 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ovn-combined-ca-bundle\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976821 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-telemetry-combined-ca-bundle\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976857 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.976905 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-libvirt-combined-ca-bundle\") pod \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\" (UID: \"d2e22e53-6826-4a58-abb0-eb2d279d2b1c\") " Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.983876 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.987856 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.987941 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.988132 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.988812 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.989083 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.990337 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.990993 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.991504 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.992043 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-kube-api-access-bg88m" (OuterVolumeSpecName: "kube-api-access-bg88m") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "kube-api-access-bg88m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:37:31 crc kubenswrapper[4810]: I0930 08:37:31.997080 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.000643 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.021888 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-inventory" (OuterVolumeSpecName: "inventory") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.046669 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2e22e53-6826-4a58-abb0-eb2d279d2b1c" (UID: "d2e22e53-6826-4a58-abb0-eb2d279d2b1c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078684 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078721 4810 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078735 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg88m\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-kube-api-access-bg88m\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078745 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078755 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078764 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078774 4810 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078787 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078795 4810 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078804 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078816 4810 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078826 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078838 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.078851 4810 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e22e53-6826-4a58-abb0-eb2d279d2b1c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.314677 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" event={"ID":"d2e22e53-6826-4a58-abb0-eb2d279d2b1c","Type":"ContainerDied","Data":"a24f250f208f11051162908847497fa896aa3c6ee98fd3689858cf121bcf6570"} Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.314991 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a24f250f208f11051162908847497fa896aa3c6ee98fd3689858cf121bcf6570" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.314779 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2rltf" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.420727 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r"] Sep 30 08:37:32 crc kubenswrapper[4810]: E0930 08:37:32.421213 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e22e53-6826-4a58-abb0-eb2d279d2b1c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.421240 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e22e53-6826-4a58-abb0-eb2d279d2b1c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.421511 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e22e53-6826-4a58-abb0-eb2d279d2b1c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.424869 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.427377 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.427426 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.427490 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.427627 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.427801 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.438439 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r"] Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.597814 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.597883 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.597906 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpc99\" (UniqueName: \"kubernetes.io/projected/dbee734f-4300-41cc-9bb9-fa005daa1df0-kube-api-access-qpc99\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.598186 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.598525 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.700150 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.700233 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.700295 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.700326 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpc99\" (UniqueName: \"kubernetes.io/projected/dbee734f-4300-41cc-9bb9-fa005daa1df0-kube-api-access-qpc99\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.700416 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.702020 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.705327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.709928 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.710880 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.723091 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpc99\" (UniqueName: \"kubernetes.io/projected/dbee734f-4300-41cc-9bb9-fa005daa1df0-kube-api-access-qpc99\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5xm2r\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.747405 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.995373 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:32 crc kubenswrapper[4810]: I0930 08:37:32.996799 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:33 crc kubenswrapper[4810]: I0930 08:37:33.463171 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r"] Sep 30 08:37:33 crc kubenswrapper[4810]: W0930 08:37:33.465024 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbee734f_4300_41cc_9bb9_fa005daa1df0.slice/crio-1e14cf82d95499ee83845220209bcda867a90a4d661e6c7c348c0451a7258239 WatchSource:0}: Error finding container 1e14cf82d95499ee83845220209bcda867a90a4d661e6c7c348c0451a7258239: Status 404 returned error can't find the container with id 1e14cf82d95499ee83845220209bcda867a90a4d661e6c7c348c0451a7258239 Sep 30 08:37:34 crc kubenswrapper[4810]: I0930 08:37:34.059225 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gd9rg" podUID="fa76491c-76b2-446b-b46b-489fb26c062e" containerName="registry-server" probeResult="failure" output=< Sep 30 08:37:34 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Sep 30 08:37:34 crc kubenswrapper[4810]: > Sep 30 08:37:34 crc kubenswrapper[4810]: I0930 08:37:34.339704 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" event={"ID":"dbee734f-4300-41cc-9bb9-fa005daa1df0","Type":"ContainerStarted","Data":"1e14cf82d95499ee83845220209bcda867a90a4d661e6c7c348c0451a7258239"} Sep 30 08:37:35 crc kubenswrapper[4810]: I0930 08:37:35.350237 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" event={"ID":"dbee734f-4300-41cc-9bb9-fa005daa1df0","Type":"ContainerStarted","Data":"3a483b802d81c91b211bad319cb3410a993753f5bf68415f719fde307cd99f50"} Sep 30 08:37:35 crc kubenswrapper[4810]: I0930 08:37:35.372657 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" podStartSLOduration=1.9321339069999999 podStartE2EDuration="3.372633333s" podCreationTimestamp="2025-09-30 08:37:32 +0000 UTC" firstStartedPulling="2025-09-30 08:37:33.467335926 +0000 UTC m=+2076.919535193" lastFinishedPulling="2025-09-30 08:37:34.907835352 +0000 UTC m=+2078.360034619" observedRunningTime="2025-09-30 08:37:35.36480807 +0000 UTC m=+2078.817007337" watchObservedRunningTime="2025-09-30 08:37:35.372633333 +0000 UTC m=+2078.824832610" Sep 30 08:37:43 crc kubenswrapper[4810]: I0930 08:37:43.052481 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:43 crc kubenswrapper[4810]: I0930 08:37:43.122203 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:43 crc kubenswrapper[4810]: I0930 08:37:43.300416 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gd9rg"] Sep 30 08:37:44 crc kubenswrapper[4810]: I0930 08:37:44.437432 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gd9rg" podUID="fa76491c-76b2-446b-b46b-489fb26c062e" containerName="registry-server" containerID="cri-o://d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66" gracePeriod=2 Sep 30 08:37:44 crc kubenswrapper[4810]: I0930 08:37:44.942179 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:44.996129 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-utilities\") pod \"fa76491c-76b2-446b-b46b-489fb26c062e\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:44.996190 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwvfs\" (UniqueName: \"kubernetes.io/projected/fa76491c-76b2-446b-b46b-489fb26c062e-kube-api-access-hwvfs\") pod \"fa76491c-76b2-446b-b46b-489fb26c062e\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:44.996249 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-catalog-content\") pod \"fa76491c-76b2-446b-b46b-489fb26c062e\" (UID: \"fa76491c-76b2-446b-b46b-489fb26c062e\") " Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:44.997576 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-utilities" (OuterVolumeSpecName: "utilities") pod "fa76491c-76b2-446b-b46b-489fb26c062e" (UID: "fa76491c-76b2-446b-b46b-489fb26c062e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.009474 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa76491c-76b2-446b-b46b-489fb26c062e-kube-api-access-hwvfs" (OuterVolumeSpecName: "kube-api-access-hwvfs") pod "fa76491c-76b2-446b-b46b-489fb26c062e" (UID: "fa76491c-76b2-446b-b46b-489fb26c062e"). InnerVolumeSpecName "kube-api-access-hwvfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.097742 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa76491c-76b2-446b-b46b-489fb26c062e" (UID: "fa76491c-76b2-446b-b46b-489fb26c062e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.099026 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.099041 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwvfs\" (UniqueName: \"kubernetes.io/projected/fa76491c-76b2-446b-b46b-489fb26c062e-kube-api-access-hwvfs\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.099050 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa76491c-76b2-446b-b46b-489fb26c062e-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.448664 4810 generic.go:334] "Generic (PLEG): container finished" podID="fa76491c-76b2-446b-b46b-489fb26c062e" containerID="d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66" exitCode=0 Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.448709 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd9rg" event={"ID":"fa76491c-76b2-446b-b46b-489fb26c062e","Type":"ContainerDied","Data":"d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66"} Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.448737 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd9rg" event={"ID":"fa76491c-76b2-446b-b46b-489fb26c062e","Type":"ContainerDied","Data":"2cbe6f46ae9c93b7a7da09599462c520259bf2cb3d6567161a4347d507697bc6"} Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.448740 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gd9rg" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.448770 4810 scope.go:117] "RemoveContainer" containerID="d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.505983 4810 scope.go:117] "RemoveContainer" containerID="ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.510454 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gd9rg"] Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.518893 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gd9rg"] Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.525409 4810 scope.go:117] "RemoveContainer" containerID="747ffe1618e805045cb758cf85a57e90d9726c2ea8a46d6733108913bdcd932a" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.568141 4810 scope.go:117] "RemoveContainer" containerID="d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66" Sep 30 08:37:45 crc kubenswrapper[4810]: E0930 08:37:45.569024 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66\": container with ID starting with d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66 not found: ID does not exist" containerID="d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.569288 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66"} err="failed to get container status \"d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66\": rpc error: code = NotFound desc = could not find container \"d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66\": container with ID starting with d2a7f269f789b93c21b891a63c7c06b83c020a96339093ac9b48d44561ab5e66 not found: ID does not exist" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.569533 4810 scope.go:117] "RemoveContainer" containerID="ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7" Sep 30 08:37:45 crc kubenswrapper[4810]: E0930 08:37:45.570118 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7\": container with ID starting with ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7 not found: ID does not exist" containerID="ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.570154 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7"} err="failed to get container status \"ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7\": rpc error: code = NotFound desc = could not find container \"ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7\": container with ID starting with ec1ec0e47c70045cce4ba3e8ebc7caef12d372b330bc999bd44d86b63b4360b7 not found: ID does not exist" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.570175 4810 scope.go:117] "RemoveContainer" containerID="747ffe1618e805045cb758cf85a57e90d9726c2ea8a46d6733108913bdcd932a" Sep 30 08:37:45 crc kubenswrapper[4810]: E0930 08:37:45.570750 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"747ffe1618e805045cb758cf85a57e90d9726c2ea8a46d6733108913bdcd932a\": container with ID starting with 747ffe1618e805045cb758cf85a57e90d9726c2ea8a46d6733108913bdcd932a not found: ID does not exist" containerID="747ffe1618e805045cb758cf85a57e90d9726c2ea8a46d6733108913bdcd932a" Sep 30 08:37:45 crc kubenswrapper[4810]: I0930 08:37:45.570997 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"747ffe1618e805045cb758cf85a57e90d9726c2ea8a46d6733108913bdcd932a"} err="failed to get container status \"747ffe1618e805045cb758cf85a57e90d9726c2ea8a46d6733108913bdcd932a\": rpc error: code = NotFound desc = could not find container \"747ffe1618e805045cb758cf85a57e90d9726c2ea8a46d6733108913bdcd932a\": container with ID starting with 747ffe1618e805045cb758cf85a57e90d9726c2ea8a46d6733108913bdcd932a not found: ID does not exist" Sep 30 08:37:47 crc kubenswrapper[4810]: I0930 08:37:47.316933 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa76491c-76b2-446b-b46b-489fb26c062e" path="/var/lib/kubelet/pods/fa76491c-76b2-446b-b46b-489fb26c062e/volumes" Sep 30 08:38:45 crc kubenswrapper[4810]: I0930 08:38:45.911940 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:38:45 crc kubenswrapper[4810]: I0930 08:38:45.912628 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:38:50 crc kubenswrapper[4810]: I0930 08:38:50.206545 4810 generic.go:334] "Generic (PLEG): container finished" podID="dbee734f-4300-41cc-9bb9-fa005daa1df0" containerID="3a483b802d81c91b211bad319cb3410a993753f5bf68415f719fde307cd99f50" exitCode=0 Sep 30 08:38:50 crc kubenswrapper[4810]: I0930 08:38:50.206656 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" event={"ID":"dbee734f-4300-41cc-9bb9-fa005daa1df0","Type":"ContainerDied","Data":"3a483b802d81c91b211bad319cb3410a993753f5bf68415f719fde307cd99f50"} Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.728185 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.753894 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovn-combined-ca-bundle\") pod \"dbee734f-4300-41cc-9bb9-fa005daa1df0\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.754028 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpc99\" (UniqueName: \"kubernetes.io/projected/dbee734f-4300-41cc-9bb9-fa005daa1df0-kube-api-access-qpc99\") pod \"dbee734f-4300-41cc-9bb9-fa005daa1df0\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.754131 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ssh-key\") pod \"dbee734f-4300-41cc-9bb9-fa005daa1df0\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.754204 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-inventory\") pod \"dbee734f-4300-41cc-9bb9-fa005daa1df0\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.754232 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovncontroller-config-0\") pod \"dbee734f-4300-41cc-9bb9-fa005daa1df0\" (UID: \"dbee734f-4300-41cc-9bb9-fa005daa1df0\") " Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.770773 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbee734f-4300-41cc-9bb9-fa005daa1df0-kube-api-access-qpc99" (OuterVolumeSpecName: "kube-api-access-qpc99") pod "dbee734f-4300-41cc-9bb9-fa005daa1df0" (UID: "dbee734f-4300-41cc-9bb9-fa005daa1df0"). InnerVolumeSpecName "kube-api-access-qpc99". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.778504 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "dbee734f-4300-41cc-9bb9-fa005daa1df0" (UID: "dbee734f-4300-41cc-9bb9-fa005daa1df0"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.786889 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-inventory" (OuterVolumeSpecName: "inventory") pod "dbee734f-4300-41cc-9bb9-fa005daa1df0" (UID: "dbee734f-4300-41cc-9bb9-fa005daa1df0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.790885 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dbee734f-4300-41cc-9bb9-fa005daa1df0" (UID: "dbee734f-4300-41cc-9bb9-fa005daa1df0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.793349 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "dbee734f-4300-41cc-9bb9-fa005daa1df0" (UID: "dbee734f-4300-41cc-9bb9-fa005daa1df0"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.856410 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.856632 4810 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.856758 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.856843 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpc99\" (UniqueName: \"kubernetes.io/projected/dbee734f-4300-41cc-9bb9-fa005daa1df0-kube-api-access-qpc99\") on node \"crc\" DevicePath \"\"" Sep 30 08:38:51 crc kubenswrapper[4810]: I0930 08:38:51.856928 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbee734f-4300-41cc-9bb9-fa005daa1df0-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.231248 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" event={"ID":"dbee734f-4300-41cc-9bb9-fa005daa1df0","Type":"ContainerDied","Data":"1e14cf82d95499ee83845220209bcda867a90a4d661e6c7c348c0451a7258239"} Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.231598 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e14cf82d95499ee83845220209bcda867a90a4d661e6c7c348c0451a7258239" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.231351 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5xm2r" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.347310 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z"] Sep 30 08:38:52 crc kubenswrapper[4810]: E0930 08:38:52.347780 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa76491c-76b2-446b-b46b-489fb26c062e" containerName="registry-server" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.347799 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa76491c-76b2-446b-b46b-489fb26c062e" containerName="registry-server" Sep 30 08:38:52 crc kubenswrapper[4810]: E0930 08:38:52.347827 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbee734f-4300-41cc-9bb9-fa005daa1df0" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.347840 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbee734f-4300-41cc-9bb9-fa005daa1df0" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 08:38:52 crc kubenswrapper[4810]: E0930 08:38:52.347857 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa76491c-76b2-446b-b46b-489fb26c062e" containerName="extract-utilities" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.347867 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa76491c-76b2-446b-b46b-489fb26c062e" containerName="extract-utilities" Sep 30 08:38:52 crc kubenswrapper[4810]: E0930 08:38:52.347906 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa76491c-76b2-446b-b46b-489fb26c062e" containerName="extract-content" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.347914 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa76491c-76b2-446b-b46b-489fb26c062e" containerName="extract-content" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.348125 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbee734f-4300-41cc-9bb9-fa005daa1df0" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.348157 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa76491c-76b2-446b-b46b-489fb26c062e" containerName="registry-server" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.349003 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.353143 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.353232 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.353143 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.353334 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.353538 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.353974 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.362889 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z"] Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.368349 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.368406 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc47t\" (UniqueName: \"kubernetes.io/projected/5de6b629-1489-4099-89bf-7773b0bdaeb8-kube-api-access-pc47t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.368488 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.368653 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.368847 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.368900 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.470781 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.470899 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.470940 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.471036 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.471069 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc47t\" (UniqueName: \"kubernetes.io/projected/5de6b629-1489-4099-89bf-7773b0bdaeb8-kube-api-access-pc47t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.471104 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.477237 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.478047 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.484078 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.487829 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.490961 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.497130 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc47t\" (UniqueName: \"kubernetes.io/projected/5de6b629-1489-4099-89bf-7773b0bdaeb8-kube-api-access-pc47t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:52 crc kubenswrapper[4810]: I0930 08:38:52.675268 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:38:53 crc kubenswrapper[4810]: I0930 08:38:53.240089 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z"] Sep 30 08:38:53 crc kubenswrapper[4810]: I0930 08:38:53.242061 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 08:38:54 crc kubenswrapper[4810]: I0930 08:38:54.264130 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" event={"ID":"5de6b629-1489-4099-89bf-7773b0bdaeb8","Type":"ContainerStarted","Data":"d5ad33af9c4f76ba14ef1807b7c1c8e59135cb8d8494a8861ed176edd5d8bede"} Sep 30 08:38:55 crc kubenswrapper[4810]: I0930 08:38:55.281503 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" event={"ID":"5de6b629-1489-4099-89bf-7773b0bdaeb8","Type":"ContainerStarted","Data":"2244de3d6adc2bbe3ecc515413c27e574e902b7bbd3ca7b9389e6f51bb7ec071"} Sep 30 08:38:55 crc kubenswrapper[4810]: I0930 08:38:55.327237 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" podStartSLOduration=2.525612616 podStartE2EDuration="3.327211886s" podCreationTimestamp="2025-09-30 08:38:52 +0000 UTC" firstStartedPulling="2025-09-30 08:38:53.241782195 +0000 UTC m=+2156.693981492" lastFinishedPulling="2025-09-30 08:38:54.043381455 +0000 UTC m=+2157.495580762" observedRunningTime="2025-09-30 08:38:55.326092773 +0000 UTC m=+2158.778292040" watchObservedRunningTime="2025-09-30 08:38:55.327211886 +0000 UTC m=+2158.779411193" Sep 30 08:39:15 crc kubenswrapper[4810]: I0930 08:39:15.912426 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:39:15 crc kubenswrapper[4810]: I0930 08:39:15.913069 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:39:45 crc kubenswrapper[4810]: I0930 08:39:45.911844 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:39:45 crc kubenswrapper[4810]: I0930 08:39:45.912486 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:39:45 crc kubenswrapper[4810]: I0930 08:39:45.912549 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:39:45 crc kubenswrapper[4810]: I0930 08:39:45.913596 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"04dd0426284d3a6d1a85a59da9e842cfb65296cf5e82dcf8080e9857a5eb8760"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:39:45 crc kubenswrapper[4810]: I0930 08:39:45.913692 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://04dd0426284d3a6d1a85a59da9e842cfb65296cf5e82dcf8080e9857a5eb8760" gracePeriod=600 Sep 30 08:39:46 crc kubenswrapper[4810]: I0930 08:39:46.914661 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="04dd0426284d3a6d1a85a59da9e842cfb65296cf5e82dcf8080e9857a5eb8760" exitCode=0 Sep 30 08:39:46 crc kubenswrapper[4810]: I0930 08:39:46.914724 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"04dd0426284d3a6d1a85a59da9e842cfb65296cf5e82dcf8080e9857a5eb8760"} Sep 30 08:39:46 crc kubenswrapper[4810]: I0930 08:39:46.915437 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d"} Sep 30 08:39:46 crc kubenswrapper[4810]: I0930 08:39:46.915463 4810 scope.go:117] "RemoveContainer" containerID="8d463c96f9742b397a808bdbc03b7079a5dd87ba2d0405630bc806fe02e81f54" Sep 30 08:39:50 crc kubenswrapper[4810]: I0930 08:39:50.964363 4810 generic.go:334] "Generic (PLEG): container finished" podID="5de6b629-1489-4099-89bf-7773b0bdaeb8" containerID="2244de3d6adc2bbe3ecc515413c27e574e902b7bbd3ca7b9389e6f51bb7ec071" exitCode=0 Sep 30 08:39:50 crc kubenswrapper[4810]: I0930 08:39:50.966138 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" event={"ID":"5de6b629-1489-4099-89bf-7773b0bdaeb8","Type":"ContainerDied","Data":"2244de3d6adc2bbe3ecc515413c27e574e902b7bbd3ca7b9389e6f51bb7ec071"} Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.480953 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.624434 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc47t\" (UniqueName: \"kubernetes.io/projected/5de6b629-1489-4099-89bf-7773b0bdaeb8-kube-api-access-pc47t\") pod \"5de6b629-1489-4099-89bf-7773b0bdaeb8\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.624610 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-ssh-key\") pod \"5de6b629-1489-4099-89bf-7773b0bdaeb8\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.624744 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-nova-metadata-neutron-config-0\") pod \"5de6b629-1489-4099-89bf-7773b0bdaeb8\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.624812 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-inventory\") pod \"5de6b629-1489-4099-89bf-7773b0bdaeb8\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.624880 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-metadata-combined-ca-bundle\") pod \"5de6b629-1489-4099-89bf-7773b0bdaeb8\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.624906 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"5de6b629-1489-4099-89bf-7773b0bdaeb8\" (UID: \"5de6b629-1489-4099-89bf-7773b0bdaeb8\") " Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.635501 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5de6b629-1489-4099-89bf-7773b0bdaeb8" (UID: "5de6b629-1489-4099-89bf-7773b0bdaeb8"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.639072 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de6b629-1489-4099-89bf-7773b0bdaeb8-kube-api-access-pc47t" (OuterVolumeSpecName: "kube-api-access-pc47t") pod "5de6b629-1489-4099-89bf-7773b0bdaeb8" (UID: "5de6b629-1489-4099-89bf-7773b0bdaeb8"). InnerVolumeSpecName "kube-api-access-pc47t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.665822 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5de6b629-1489-4099-89bf-7773b0bdaeb8" (UID: "5de6b629-1489-4099-89bf-7773b0bdaeb8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.682404 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "5de6b629-1489-4099-89bf-7773b0bdaeb8" (UID: "5de6b629-1489-4099-89bf-7773b0bdaeb8"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.683676 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "5de6b629-1489-4099-89bf-7773b0bdaeb8" (UID: "5de6b629-1489-4099-89bf-7773b0bdaeb8"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.685699 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-inventory" (OuterVolumeSpecName: "inventory") pod "5de6b629-1489-4099-89bf-7773b0bdaeb8" (UID: "5de6b629-1489-4099-89bf-7773b0bdaeb8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.727740 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.727782 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.727802 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.727816 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc47t\" (UniqueName: \"kubernetes.io/projected/5de6b629-1489-4099-89bf-7773b0bdaeb8-kube-api-access-pc47t\") on node \"crc\" DevicePath \"\"" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.727828 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:39:52 crc kubenswrapper[4810]: I0930 08:39:52.727840 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5de6b629-1489-4099-89bf-7773b0bdaeb8-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.001047 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" event={"ID":"5de6b629-1489-4099-89bf-7773b0bdaeb8","Type":"ContainerDied","Data":"d5ad33af9c4f76ba14ef1807b7c1c8e59135cb8d8494a8861ed176edd5d8bede"} Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.001352 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5ad33af9c4f76ba14ef1807b7c1c8e59135cb8d8494a8861ed176edd5d8bede" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.001104 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.138371 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd"] Sep 30 08:39:53 crc kubenswrapper[4810]: E0930 08:39:53.138801 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de6b629-1489-4099-89bf-7773b0bdaeb8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.138820 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de6b629-1489-4099-89bf-7773b0bdaeb8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.138995 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de6b629-1489-4099-89bf-7773b0bdaeb8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.139691 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.143178 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.143656 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.143848 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.143979 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.145284 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.161144 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd"] Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.238881 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99mwt\" (UniqueName: \"kubernetes.io/projected/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-kube-api-access-99mwt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.239019 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.239090 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.239146 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.239538 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.341377 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.341461 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.341521 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.341582 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.341648 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99mwt\" (UniqueName: \"kubernetes.io/projected/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-kube-api-access-99mwt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.350120 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.350989 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.352219 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.354906 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.363098 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99mwt\" (UniqueName: \"kubernetes.io/projected/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-kube-api-access-99mwt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:53 crc kubenswrapper[4810]: I0930 08:39:53.466825 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:39:54 crc kubenswrapper[4810]: I0930 08:39:54.180338 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd"] Sep 30 08:39:55 crc kubenswrapper[4810]: I0930 08:39:55.027834 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" event={"ID":"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1","Type":"ContainerStarted","Data":"a4015b2121f31509fb396fef839c6ee027acb809dba25438a1c30690fe6ead9c"} Sep 30 08:39:55 crc kubenswrapper[4810]: I0930 08:39:55.028206 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" event={"ID":"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1","Type":"ContainerStarted","Data":"f3a0106e0e485808783d9fa501415ac208f6a730706673052299e5d5af8e8614"} Sep 30 08:39:55 crc kubenswrapper[4810]: I0930 08:39:55.051033 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" podStartSLOduration=1.5247825069999998 podStartE2EDuration="2.050990436s" podCreationTimestamp="2025-09-30 08:39:53 +0000 UTC" firstStartedPulling="2025-09-30 08:39:54.177761293 +0000 UTC m=+2217.629960610" lastFinishedPulling="2025-09-30 08:39:54.703969242 +0000 UTC m=+2218.156168539" observedRunningTime="2025-09-30 08:39:55.044656007 +0000 UTC m=+2218.496855274" watchObservedRunningTime="2025-09-30 08:39:55.050990436 +0000 UTC m=+2218.503189743" Sep 30 08:42:15 crc kubenswrapper[4810]: I0930 08:42:15.912018 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:42:15 crc kubenswrapper[4810]: I0930 08:42:15.912704 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:42:45 crc kubenswrapper[4810]: I0930 08:42:45.911958 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:42:45 crc kubenswrapper[4810]: I0930 08:42:45.913036 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:43:15 crc kubenswrapper[4810]: I0930 08:43:15.911513 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:43:15 crc kubenswrapper[4810]: I0930 08:43:15.911960 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:43:15 crc kubenswrapper[4810]: I0930 08:43:15.912003 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:43:15 crc kubenswrapper[4810]: I0930 08:43:15.912709 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:43:15 crc kubenswrapper[4810]: I0930 08:43:15.912763 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" gracePeriod=600 Sep 30 08:43:16 crc kubenswrapper[4810]: E0930 08:43:16.045950 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:43:16 crc kubenswrapper[4810]: I0930 08:43:16.407177 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" exitCode=0 Sep 30 08:43:16 crc kubenswrapper[4810]: I0930 08:43:16.407251 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d"} Sep 30 08:43:16 crc kubenswrapper[4810]: I0930 08:43:16.407362 4810 scope.go:117] "RemoveContainer" containerID="04dd0426284d3a6d1a85a59da9e842cfb65296cf5e82dcf8080e9857a5eb8760" Sep 30 08:43:16 crc kubenswrapper[4810]: I0930 08:43:16.408368 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:43:16 crc kubenswrapper[4810]: E0930 08:43:16.409101 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:43:28 crc kubenswrapper[4810]: I0930 08:43:28.306789 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:43:28 crc kubenswrapper[4810]: E0930 08:43:28.307766 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:43:40 crc kubenswrapper[4810]: I0930 08:43:40.309358 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:43:40 crc kubenswrapper[4810]: E0930 08:43:40.309958 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:43:51 crc kubenswrapper[4810]: I0930 08:43:51.306835 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:43:51 crc kubenswrapper[4810]: E0930 08:43:51.307539 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:44:06 crc kubenswrapper[4810]: I0930 08:44:06.307708 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:44:06 crc kubenswrapper[4810]: E0930 08:44:06.308800 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:44:17 crc kubenswrapper[4810]: I0930 08:44:17.323701 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:44:17 crc kubenswrapper[4810]: E0930 08:44:17.324893 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:44:29 crc kubenswrapper[4810]: I0930 08:44:29.306511 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:44:29 crc kubenswrapper[4810]: E0930 08:44:29.307289 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:44:40 crc kubenswrapper[4810]: I0930 08:44:40.410925 4810 generic.go:334] "Generic (PLEG): container finished" podID="9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1" containerID="a4015b2121f31509fb396fef839c6ee027acb809dba25438a1c30690fe6ead9c" exitCode=0 Sep 30 08:44:40 crc kubenswrapper[4810]: I0930 08:44:40.411067 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" event={"ID":"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1","Type":"ContainerDied","Data":"a4015b2121f31509fb396fef839c6ee027acb809dba25438a1c30690fe6ead9c"} Sep 30 08:44:41 crc kubenswrapper[4810]: I0930 08:44:41.881772 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.011412 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-secret-0\") pod \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.011512 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99mwt\" (UniqueName: \"kubernetes.io/projected/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-kube-api-access-99mwt\") pod \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.011554 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-combined-ca-bundle\") pod \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.011578 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-inventory\") pod \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.011604 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-ssh-key\") pod \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\" (UID: \"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1\") " Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.017187 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1" (UID: "9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.022983 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-kube-api-access-99mwt" (OuterVolumeSpecName: "kube-api-access-99mwt") pod "9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1" (UID: "9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1"). InnerVolumeSpecName "kube-api-access-99mwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.047416 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-inventory" (OuterVolumeSpecName: "inventory") pod "9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1" (UID: "9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.052798 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1" (UID: "9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.077945 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1" (UID: "9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.114365 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.114405 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.114418 4810 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.114605 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99mwt\" (UniqueName: \"kubernetes.io/projected/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-kube-api-access-99mwt\") on node \"crc\" DevicePath \"\"" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.114623 4810 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.437580 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" event={"ID":"9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1","Type":"ContainerDied","Data":"f3a0106e0e485808783d9fa501415ac208f6a730706673052299e5d5af8e8614"} Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.437648 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3a0106e0e485808783d9fa501415ac208f6a730706673052299e5d5af8e8614" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.437743 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.578239 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68"] Sep 30 08:44:42 crc kubenswrapper[4810]: E0930 08:44:42.579008 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.579031 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.579537 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.580596 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.583607 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.583896 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.584109 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.584246 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.584418 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.584536 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.584669 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.607551 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68"] Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.725184 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.725606 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.725657 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.725714 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.725737 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.725786 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.725809 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.725844 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdhpl\" (UniqueName: \"kubernetes.io/projected/8a25c107-d531-4a79-97e9-7201d6a8485a-kube-api-access-pdhpl\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.726044 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.828975 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.829172 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.829216 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.829320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.829362 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.829454 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdhpl\" (UniqueName: \"kubernetes.io/projected/8a25c107-d531-4a79-97e9-7201d6a8485a-kube-api-access-pdhpl\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.829692 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.829782 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.829918 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.830333 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.833694 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.833707 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.835218 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.835311 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.835892 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.837187 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.844327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.849770 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdhpl\" (UniqueName: \"kubernetes.io/projected/8a25c107-d531-4a79-97e9-7201d6a8485a-kube-api-access-pdhpl\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v6q68\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:42 crc kubenswrapper[4810]: I0930 08:44:42.912724 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:44:43 crc kubenswrapper[4810]: I0930 08:44:43.500533 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68"] Sep 30 08:44:43 crc kubenswrapper[4810]: W0930 08:44:43.502358 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a25c107_d531_4a79_97e9_7201d6a8485a.slice/crio-4033261d3c6d74bee294b2cd87b9613c94992498d1cb2b09a8dc556e9efb5cd2 WatchSource:0}: Error finding container 4033261d3c6d74bee294b2cd87b9613c94992498d1cb2b09a8dc556e9efb5cd2: Status 404 returned error can't find the container with id 4033261d3c6d74bee294b2cd87b9613c94992498d1cb2b09a8dc556e9efb5cd2 Sep 30 08:44:43 crc kubenswrapper[4810]: I0930 08:44:43.505537 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 08:44:44 crc kubenswrapper[4810]: I0930 08:44:44.306591 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:44:44 crc kubenswrapper[4810]: E0930 08:44:44.307537 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:44:44 crc kubenswrapper[4810]: I0930 08:44:44.466706 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" event={"ID":"8a25c107-d531-4a79-97e9-7201d6a8485a","Type":"ContainerStarted","Data":"c91606951ae9cf90fb0c34347a580a07530f4b5b2062458a0481adcdcbe57e18"} Sep 30 08:44:44 crc kubenswrapper[4810]: I0930 08:44:44.466774 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" event={"ID":"8a25c107-d531-4a79-97e9-7201d6a8485a","Type":"ContainerStarted","Data":"4033261d3c6d74bee294b2cd87b9613c94992498d1cb2b09a8dc556e9efb5cd2"} Sep 30 08:44:44 crc kubenswrapper[4810]: I0930 08:44:44.492771 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" podStartSLOduration=1.98768239 podStartE2EDuration="2.492739757s" podCreationTimestamp="2025-09-30 08:44:42 +0000 UTC" firstStartedPulling="2025-09-30 08:44:43.504940152 +0000 UTC m=+2506.957139459" lastFinishedPulling="2025-09-30 08:44:44.009997529 +0000 UTC m=+2507.462196826" observedRunningTime="2025-09-30 08:44:44.48697656 +0000 UTC m=+2507.939175857" watchObservedRunningTime="2025-09-30 08:44:44.492739757 +0000 UTC m=+2507.944939064" Sep 30 08:44:59 crc kubenswrapper[4810]: I0930 08:44:59.307493 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:44:59 crc kubenswrapper[4810]: E0930 08:44:59.308638 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.153872 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s"] Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.156552 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.163050 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.163357 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.181810 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s"] Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.221822 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/07460fac-bba2-429f-b0f0-5e569719e887-config-volume\") pod \"collect-profiles-29320365-q545s\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.222232 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snf9t\" (UniqueName: \"kubernetes.io/projected/07460fac-bba2-429f-b0f0-5e569719e887-kube-api-access-snf9t\") pod \"collect-profiles-29320365-q545s\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.222466 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/07460fac-bba2-429f-b0f0-5e569719e887-secret-volume\") pod \"collect-profiles-29320365-q545s\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.323722 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/07460fac-bba2-429f-b0f0-5e569719e887-secret-volume\") pod \"collect-profiles-29320365-q545s\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.324295 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/07460fac-bba2-429f-b0f0-5e569719e887-config-volume\") pod \"collect-profiles-29320365-q545s\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.324451 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snf9t\" (UniqueName: \"kubernetes.io/projected/07460fac-bba2-429f-b0f0-5e569719e887-kube-api-access-snf9t\") pod \"collect-profiles-29320365-q545s\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.325151 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/07460fac-bba2-429f-b0f0-5e569719e887-config-volume\") pod \"collect-profiles-29320365-q545s\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.335347 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/07460fac-bba2-429f-b0f0-5e569719e887-secret-volume\") pod \"collect-profiles-29320365-q545s\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.340137 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snf9t\" (UniqueName: \"kubernetes.io/projected/07460fac-bba2-429f-b0f0-5e569719e887-kube-api-access-snf9t\") pod \"collect-profiles-29320365-q545s\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.494368 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:00 crc kubenswrapper[4810]: I0930 08:45:00.974354 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s"] Sep 30 08:45:01 crc kubenswrapper[4810]: I0930 08:45:01.661989 4810 generic.go:334] "Generic (PLEG): container finished" podID="07460fac-bba2-429f-b0f0-5e569719e887" containerID="3aaea24f70ebaa8415966583515ecaa5dc71ef91585138b665cafe07b2c4396b" exitCode=0 Sep 30 08:45:01 crc kubenswrapper[4810]: I0930 08:45:01.662093 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" event={"ID":"07460fac-bba2-429f-b0f0-5e569719e887","Type":"ContainerDied","Data":"3aaea24f70ebaa8415966583515ecaa5dc71ef91585138b665cafe07b2c4396b"} Sep 30 08:45:01 crc kubenswrapper[4810]: I0930 08:45:01.662444 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" event={"ID":"07460fac-bba2-429f-b0f0-5e569719e887","Type":"ContainerStarted","Data":"959b0a40591008039f052d4e21d672312275f9930a322885ca012423fb16ed53"} Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.059901 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.190131 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/07460fac-bba2-429f-b0f0-5e569719e887-config-volume\") pod \"07460fac-bba2-429f-b0f0-5e569719e887\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.190629 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/07460fac-bba2-429f-b0f0-5e569719e887-secret-volume\") pod \"07460fac-bba2-429f-b0f0-5e569719e887\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.190824 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snf9t\" (UniqueName: \"kubernetes.io/projected/07460fac-bba2-429f-b0f0-5e569719e887-kube-api-access-snf9t\") pod \"07460fac-bba2-429f-b0f0-5e569719e887\" (UID: \"07460fac-bba2-429f-b0f0-5e569719e887\") " Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.191693 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07460fac-bba2-429f-b0f0-5e569719e887-config-volume" (OuterVolumeSpecName: "config-volume") pod "07460fac-bba2-429f-b0f0-5e569719e887" (UID: "07460fac-bba2-429f-b0f0-5e569719e887"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.206335 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07460fac-bba2-429f-b0f0-5e569719e887-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "07460fac-bba2-429f-b0f0-5e569719e887" (UID: "07460fac-bba2-429f-b0f0-5e569719e887"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.206574 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07460fac-bba2-429f-b0f0-5e569719e887-kube-api-access-snf9t" (OuterVolumeSpecName: "kube-api-access-snf9t") pod "07460fac-bba2-429f-b0f0-5e569719e887" (UID: "07460fac-bba2-429f-b0f0-5e569719e887"). InnerVolumeSpecName "kube-api-access-snf9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.293241 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/07460fac-bba2-429f-b0f0-5e569719e887-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.293310 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/07460fac-bba2-429f-b0f0-5e569719e887-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.293324 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snf9t\" (UniqueName: \"kubernetes.io/projected/07460fac-bba2-429f-b0f0-5e569719e887-kube-api-access-snf9t\") on node \"crc\" DevicePath \"\"" Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.687757 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" event={"ID":"07460fac-bba2-429f-b0f0-5e569719e887","Type":"ContainerDied","Data":"959b0a40591008039f052d4e21d672312275f9930a322885ca012423fb16ed53"} Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.687819 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s" Sep 30 08:45:03 crc kubenswrapper[4810]: I0930 08:45:03.687823 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="959b0a40591008039f052d4e21d672312275f9930a322885ca012423fb16ed53" Sep 30 08:45:04 crc kubenswrapper[4810]: I0930 08:45:04.168799 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn"] Sep 30 08:45:04 crc kubenswrapper[4810]: I0930 08:45:04.176650 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320320-kzpbn"] Sep 30 08:45:05 crc kubenswrapper[4810]: I0930 08:45:05.332932 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8d5914f-a41d-4e29-bd6b-3e5050c23cd6" path="/var/lib/kubelet/pods/b8d5914f-a41d-4e29-bd6b-3e5050c23cd6/volumes" Sep 30 08:45:12 crc kubenswrapper[4810]: I0930 08:45:12.307562 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:45:12 crc kubenswrapper[4810]: E0930 08:45:12.308682 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:45:21 crc kubenswrapper[4810]: I0930 08:45:21.356766 4810 scope.go:117] "RemoveContainer" containerID="26b716174b8617ef9307046c3f0d04ab2fdc57b54a9b7691dae85ca34cd14319" Sep 30 08:45:23 crc kubenswrapper[4810]: I0930 08:45:23.307705 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:45:23 crc kubenswrapper[4810]: E0930 08:45:23.308460 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.587702 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v6l2r"] Sep 30 08:45:34 crc kubenswrapper[4810]: E0930 08:45:34.590780 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07460fac-bba2-429f-b0f0-5e569719e887" containerName="collect-profiles" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.590868 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="07460fac-bba2-429f-b0f0-5e569719e887" containerName="collect-profiles" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.591490 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="07460fac-bba2-429f-b0f0-5e569719e887" containerName="collect-profiles" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.594945 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.606496 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v6l2r"] Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.746642 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx5w2\" (UniqueName: \"kubernetes.io/projected/9526e749-783c-41ce-90a6-c4957177754e-kube-api-access-bx5w2\") pod \"redhat-marketplace-v6l2r\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.746790 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-catalog-content\") pod \"redhat-marketplace-v6l2r\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.746849 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-utilities\") pod \"redhat-marketplace-v6l2r\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.848514 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-catalog-content\") pod \"redhat-marketplace-v6l2r\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.848595 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-utilities\") pod \"redhat-marketplace-v6l2r\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.848763 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx5w2\" (UniqueName: \"kubernetes.io/projected/9526e749-783c-41ce-90a6-c4957177754e-kube-api-access-bx5w2\") pod \"redhat-marketplace-v6l2r\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.849165 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-utilities\") pod \"redhat-marketplace-v6l2r\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.849239 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-catalog-content\") pod \"redhat-marketplace-v6l2r\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.886637 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx5w2\" (UniqueName: \"kubernetes.io/projected/9526e749-783c-41ce-90a6-c4957177754e-kube-api-access-bx5w2\") pod \"redhat-marketplace-v6l2r\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:34 crc kubenswrapper[4810]: I0930 08:45:34.966057 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:35 crc kubenswrapper[4810]: I0930 08:45:35.425320 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v6l2r"] Sep 30 08:45:36 crc kubenswrapper[4810]: I0930 08:45:36.116967 4810 generic.go:334] "Generic (PLEG): container finished" podID="9526e749-783c-41ce-90a6-c4957177754e" containerID="dd0a4e171e2776d66dd7c37c5de0215cb64037dfbda2668b959635037bc9b3db" exitCode=0 Sep 30 08:45:36 crc kubenswrapper[4810]: I0930 08:45:36.117019 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v6l2r" event={"ID":"9526e749-783c-41ce-90a6-c4957177754e","Type":"ContainerDied","Data":"dd0a4e171e2776d66dd7c37c5de0215cb64037dfbda2668b959635037bc9b3db"} Sep 30 08:45:36 crc kubenswrapper[4810]: I0930 08:45:36.117050 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v6l2r" event={"ID":"9526e749-783c-41ce-90a6-c4957177754e","Type":"ContainerStarted","Data":"40e777da344ac93904d3d792a0297149e7f6dfcf4caaf57fc7295e3b9e1d5435"} Sep 30 08:45:37 crc kubenswrapper[4810]: I0930 08:45:37.319061 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:45:37 crc kubenswrapper[4810]: E0930 08:45:37.319889 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:45:39 crc kubenswrapper[4810]: I0930 08:45:39.162063 4810 generic.go:334] "Generic (PLEG): container finished" podID="9526e749-783c-41ce-90a6-c4957177754e" containerID="5bb20494d3d65be1c9e6119311b9c579a76b7713045f3228666bae64326dd520" exitCode=0 Sep 30 08:45:39 crc kubenswrapper[4810]: I0930 08:45:39.162234 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v6l2r" event={"ID":"9526e749-783c-41ce-90a6-c4957177754e","Type":"ContainerDied","Data":"5bb20494d3d65be1c9e6119311b9c579a76b7713045f3228666bae64326dd520"} Sep 30 08:45:40 crc kubenswrapper[4810]: I0930 08:45:40.176408 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v6l2r" event={"ID":"9526e749-783c-41ce-90a6-c4957177754e","Type":"ContainerStarted","Data":"e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972"} Sep 30 08:45:40 crc kubenswrapper[4810]: I0930 08:45:40.202946 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v6l2r" podStartSLOduration=2.568802251 podStartE2EDuration="6.202920167s" podCreationTimestamp="2025-09-30 08:45:34 +0000 UTC" firstStartedPulling="2025-09-30 08:45:36.122042807 +0000 UTC m=+2559.574242084" lastFinishedPulling="2025-09-30 08:45:39.756160723 +0000 UTC m=+2563.208360000" observedRunningTime="2025-09-30 08:45:40.196375567 +0000 UTC m=+2563.648574874" watchObservedRunningTime="2025-09-30 08:45:40.202920167 +0000 UTC m=+2563.655119474" Sep 30 08:45:44 crc kubenswrapper[4810]: I0930 08:45:44.967141 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:44 crc kubenswrapper[4810]: I0930 08:45:44.968487 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:45 crc kubenswrapper[4810]: I0930 08:45:45.052491 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:45 crc kubenswrapper[4810]: I0930 08:45:45.320355 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:45 crc kubenswrapper[4810]: I0930 08:45:45.403621 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v6l2r"] Sep 30 08:45:47 crc kubenswrapper[4810]: I0930 08:45:47.285457 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v6l2r" podUID="9526e749-783c-41ce-90a6-c4957177754e" containerName="registry-server" containerID="cri-o://e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972" gracePeriod=2 Sep 30 08:45:47 crc kubenswrapper[4810]: I0930 08:45:47.840781 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:47 crc kubenswrapper[4810]: I0930 08:45:47.952901 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-catalog-content\") pod \"9526e749-783c-41ce-90a6-c4957177754e\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " Sep 30 08:45:47 crc kubenswrapper[4810]: I0930 08:45:47.953098 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx5w2\" (UniqueName: \"kubernetes.io/projected/9526e749-783c-41ce-90a6-c4957177754e-kube-api-access-bx5w2\") pod \"9526e749-783c-41ce-90a6-c4957177754e\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " Sep 30 08:45:47 crc kubenswrapper[4810]: I0930 08:45:47.953196 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-utilities\") pod \"9526e749-783c-41ce-90a6-c4957177754e\" (UID: \"9526e749-783c-41ce-90a6-c4957177754e\") " Sep 30 08:45:47 crc kubenswrapper[4810]: I0930 08:45:47.954670 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-utilities" (OuterVolumeSpecName: "utilities") pod "9526e749-783c-41ce-90a6-c4957177754e" (UID: "9526e749-783c-41ce-90a6-c4957177754e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:45:47 crc kubenswrapper[4810]: I0930 08:45:47.964300 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9526e749-783c-41ce-90a6-c4957177754e-kube-api-access-bx5w2" (OuterVolumeSpecName: "kube-api-access-bx5w2") pod "9526e749-783c-41ce-90a6-c4957177754e" (UID: "9526e749-783c-41ce-90a6-c4957177754e"). InnerVolumeSpecName "kube-api-access-bx5w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:45:47 crc kubenswrapper[4810]: I0930 08:45:47.986633 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9526e749-783c-41ce-90a6-c4957177754e" (UID: "9526e749-783c-41ce-90a6-c4957177754e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.056891 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.056948 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx5w2\" (UniqueName: \"kubernetes.io/projected/9526e749-783c-41ce-90a6-c4957177754e-kube-api-access-bx5w2\") on node \"crc\" DevicePath \"\"" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.056973 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9526e749-783c-41ce-90a6-c4957177754e-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.297084 4810 generic.go:334] "Generic (PLEG): container finished" podID="9526e749-783c-41ce-90a6-c4957177754e" containerID="e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972" exitCode=0 Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.297119 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v6l2r" event={"ID":"9526e749-783c-41ce-90a6-c4957177754e","Type":"ContainerDied","Data":"e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972"} Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.297143 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v6l2r" event={"ID":"9526e749-783c-41ce-90a6-c4957177754e","Type":"ContainerDied","Data":"40e777da344ac93904d3d792a0297149e7f6dfcf4caaf57fc7295e3b9e1d5435"} Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.297158 4810 scope.go:117] "RemoveContainer" containerID="e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.297166 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v6l2r" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.331782 4810 scope.go:117] "RemoveContainer" containerID="5bb20494d3d65be1c9e6119311b9c579a76b7713045f3228666bae64326dd520" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.355005 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v6l2r"] Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.363980 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v6l2r"] Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.366598 4810 scope.go:117] "RemoveContainer" containerID="dd0a4e171e2776d66dd7c37c5de0215cb64037dfbda2668b959635037bc9b3db" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.412216 4810 scope.go:117] "RemoveContainer" containerID="e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972" Sep 30 08:45:48 crc kubenswrapper[4810]: E0930 08:45:48.412695 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972\": container with ID starting with e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972 not found: ID does not exist" containerID="e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.412738 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972"} err="failed to get container status \"e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972\": rpc error: code = NotFound desc = could not find container \"e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972\": container with ID starting with e636251536e0c25842615c022008d987c1ab393d56f53cd8151d394a3925c972 not found: ID does not exist" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.412781 4810 scope.go:117] "RemoveContainer" containerID="5bb20494d3d65be1c9e6119311b9c579a76b7713045f3228666bae64326dd520" Sep 30 08:45:48 crc kubenswrapper[4810]: E0930 08:45:48.413192 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bb20494d3d65be1c9e6119311b9c579a76b7713045f3228666bae64326dd520\": container with ID starting with 5bb20494d3d65be1c9e6119311b9c579a76b7713045f3228666bae64326dd520 not found: ID does not exist" containerID="5bb20494d3d65be1c9e6119311b9c579a76b7713045f3228666bae64326dd520" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.413244 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb20494d3d65be1c9e6119311b9c579a76b7713045f3228666bae64326dd520"} err="failed to get container status \"5bb20494d3d65be1c9e6119311b9c579a76b7713045f3228666bae64326dd520\": rpc error: code = NotFound desc = could not find container \"5bb20494d3d65be1c9e6119311b9c579a76b7713045f3228666bae64326dd520\": container with ID starting with 5bb20494d3d65be1c9e6119311b9c579a76b7713045f3228666bae64326dd520 not found: ID does not exist" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.413297 4810 scope.go:117] "RemoveContainer" containerID="dd0a4e171e2776d66dd7c37c5de0215cb64037dfbda2668b959635037bc9b3db" Sep 30 08:45:48 crc kubenswrapper[4810]: E0930 08:45:48.413608 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd0a4e171e2776d66dd7c37c5de0215cb64037dfbda2668b959635037bc9b3db\": container with ID starting with dd0a4e171e2776d66dd7c37c5de0215cb64037dfbda2668b959635037bc9b3db not found: ID does not exist" containerID="dd0a4e171e2776d66dd7c37c5de0215cb64037dfbda2668b959635037bc9b3db" Sep 30 08:45:48 crc kubenswrapper[4810]: I0930 08:45:48.413643 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd0a4e171e2776d66dd7c37c5de0215cb64037dfbda2668b959635037bc9b3db"} err="failed to get container status \"dd0a4e171e2776d66dd7c37c5de0215cb64037dfbda2668b959635037bc9b3db\": rpc error: code = NotFound desc = could not find container \"dd0a4e171e2776d66dd7c37c5de0215cb64037dfbda2668b959635037bc9b3db\": container with ID starting with dd0a4e171e2776d66dd7c37c5de0215cb64037dfbda2668b959635037bc9b3db not found: ID does not exist" Sep 30 08:45:49 crc kubenswrapper[4810]: I0930 08:45:49.319471 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9526e749-783c-41ce-90a6-c4957177754e" path="/var/lib/kubelet/pods/9526e749-783c-41ce-90a6-c4957177754e/volumes" Sep 30 08:45:51 crc kubenswrapper[4810]: I0930 08:45:51.307104 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:45:51 crc kubenswrapper[4810]: E0930 08:45:51.307782 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:46:04 crc kubenswrapper[4810]: I0930 08:46:04.306035 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:46:04 crc kubenswrapper[4810]: E0930 08:46:04.306858 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:46:15 crc kubenswrapper[4810]: I0930 08:46:15.309478 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:46:15 crc kubenswrapper[4810]: E0930 08:46:15.310448 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:46:30 crc kubenswrapper[4810]: I0930 08:46:30.306226 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:46:30 crc kubenswrapper[4810]: E0930 08:46:30.306948 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:46:45 crc kubenswrapper[4810]: I0930 08:46:45.307021 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:46:45 crc kubenswrapper[4810]: E0930 08:46:45.307768 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:46:56 crc kubenswrapper[4810]: I0930 08:46:56.307548 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:46:56 crc kubenswrapper[4810]: E0930 08:46:56.308367 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:47:07 crc kubenswrapper[4810]: I0930 08:47:07.314551 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:47:07 crc kubenswrapper[4810]: E0930 08:47:07.315489 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:47:20 crc kubenswrapper[4810]: I0930 08:47:20.307216 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:47:20 crc kubenswrapper[4810]: E0930 08:47:20.308256 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:47:32 crc kubenswrapper[4810]: I0930 08:47:32.306939 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:47:32 crc kubenswrapper[4810]: E0930 08:47:32.307907 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.065027 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6jlnq"] Sep 30 08:47:36 crc kubenswrapper[4810]: E0930 08:47:36.068228 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9526e749-783c-41ce-90a6-c4957177754e" containerName="extract-utilities" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.068258 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9526e749-783c-41ce-90a6-c4957177754e" containerName="extract-utilities" Sep 30 08:47:36 crc kubenswrapper[4810]: E0930 08:47:36.068354 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9526e749-783c-41ce-90a6-c4957177754e" containerName="registry-server" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.068395 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9526e749-783c-41ce-90a6-c4957177754e" containerName="registry-server" Sep 30 08:47:36 crc kubenswrapper[4810]: E0930 08:47:36.068427 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9526e749-783c-41ce-90a6-c4957177754e" containerName="extract-content" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.068442 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9526e749-783c-41ce-90a6-c4957177754e" containerName="extract-content" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.068831 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9526e749-783c-41ce-90a6-c4957177754e" containerName="registry-server" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.071903 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.101346 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6jlnq"] Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.218881 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-utilities\") pod \"certified-operators-6jlnq\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.218998 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-catalog-content\") pod \"certified-operators-6jlnq\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.219145 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzz4w\" (UniqueName: \"kubernetes.io/projected/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-kube-api-access-jzz4w\") pod \"certified-operators-6jlnq\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.321500 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-utilities\") pod \"certified-operators-6jlnq\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.321548 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-catalog-content\") pod \"certified-operators-6jlnq\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.321598 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzz4w\" (UniqueName: \"kubernetes.io/projected/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-kube-api-access-jzz4w\") pod \"certified-operators-6jlnq\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.322160 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-catalog-content\") pod \"certified-operators-6jlnq\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.322572 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-utilities\") pod \"certified-operators-6jlnq\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.348342 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzz4w\" (UniqueName: \"kubernetes.io/projected/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-kube-api-access-jzz4w\") pod \"certified-operators-6jlnq\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.422956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:36 crc kubenswrapper[4810]: I0930 08:47:36.963508 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6jlnq"] Sep 30 08:47:37 crc kubenswrapper[4810]: I0930 08:47:37.591151 4810 generic.go:334] "Generic (PLEG): container finished" podID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" containerID="2231513c8b90134e07c02b2e618fb90c10c3f43d14cf852985d5ab9a8ffa7ad6" exitCode=0 Sep 30 08:47:37 crc kubenswrapper[4810]: I0930 08:47:37.591235 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jlnq" event={"ID":"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c","Type":"ContainerDied","Data":"2231513c8b90134e07c02b2e618fb90c10c3f43d14cf852985d5ab9a8ffa7ad6"} Sep 30 08:47:37 crc kubenswrapper[4810]: I0930 08:47:37.591530 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jlnq" event={"ID":"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c","Type":"ContainerStarted","Data":"8e2dc4dc1ec72770fbceb9cf26127a10e36276d8e5aec31fb036659c6338b8cc"} Sep 30 08:47:38 crc kubenswrapper[4810]: I0930 08:47:38.606081 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jlnq" event={"ID":"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c","Type":"ContainerStarted","Data":"8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513"} Sep 30 08:47:39 crc kubenswrapper[4810]: I0930 08:47:39.631411 4810 generic.go:334] "Generic (PLEG): container finished" podID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" containerID="8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513" exitCode=0 Sep 30 08:47:39 crc kubenswrapper[4810]: I0930 08:47:39.631532 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jlnq" event={"ID":"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c","Type":"ContainerDied","Data":"8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513"} Sep 30 08:47:40 crc kubenswrapper[4810]: I0930 08:47:40.644457 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jlnq" event={"ID":"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c","Type":"ContainerStarted","Data":"f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c"} Sep 30 08:47:40 crc kubenswrapper[4810]: I0930 08:47:40.671497 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6jlnq" podStartSLOduration=1.978527575 podStartE2EDuration="4.67147626s" podCreationTimestamp="2025-09-30 08:47:36 +0000 UTC" firstStartedPulling="2025-09-30 08:47:37.594574135 +0000 UTC m=+2681.046773432" lastFinishedPulling="2025-09-30 08:47:40.28752281 +0000 UTC m=+2683.739722117" observedRunningTime="2025-09-30 08:47:40.664014703 +0000 UTC m=+2684.116213970" watchObservedRunningTime="2025-09-30 08:47:40.67147626 +0000 UTC m=+2684.123675537" Sep 30 08:47:45 crc kubenswrapper[4810]: I0930 08:47:45.308309 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:47:45 crc kubenswrapper[4810]: E0930 08:47:45.309225 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:47:46 crc kubenswrapper[4810]: I0930 08:47:46.423675 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:46 crc kubenswrapper[4810]: I0930 08:47:46.423739 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:46 crc kubenswrapper[4810]: I0930 08:47:46.567260 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:46 crc kubenswrapper[4810]: I0930 08:47:46.757804 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:46 crc kubenswrapper[4810]: I0930 08:47:46.799886 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6jlnq"] Sep 30 08:47:48 crc kubenswrapper[4810]: I0930 08:47:48.736166 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6jlnq" podUID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" containerName="registry-server" containerID="cri-o://f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c" gracePeriod=2 Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.249566 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.328764 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-utilities\") pod \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.329033 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzz4w\" (UniqueName: \"kubernetes.io/projected/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-kube-api-access-jzz4w\") pod \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.329078 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-catalog-content\") pod \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\" (UID: \"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c\") " Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.329893 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-utilities" (OuterVolumeSpecName: "utilities") pod "cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" (UID: "cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.338719 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-kube-api-access-jzz4w" (OuterVolumeSpecName: "kube-api-access-jzz4w") pod "cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" (UID: "cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c"). InnerVolumeSpecName "kube-api-access-jzz4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.373934 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" (UID: "cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.432630 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.432683 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.432704 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzz4w\" (UniqueName: \"kubernetes.io/projected/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c-kube-api-access-jzz4w\") on node \"crc\" DevicePath \"\"" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.751244 4810 generic.go:334] "Generic (PLEG): container finished" podID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" containerID="f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c" exitCode=0 Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.751353 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jlnq" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.751353 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jlnq" event={"ID":"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c","Type":"ContainerDied","Data":"f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c"} Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.753447 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jlnq" event={"ID":"cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c","Type":"ContainerDied","Data":"8e2dc4dc1ec72770fbceb9cf26127a10e36276d8e5aec31fb036659c6338b8cc"} Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.753487 4810 scope.go:117] "RemoveContainer" containerID="f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.778419 4810 scope.go:117] "RemoveContainer" containerID="8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.815525 4810 scope.go:117] "RemoveContainer" containerID="2231513c8b90134e07c02b2e618fb90c10c3f43d14cf852985d5ab9a8ffa7ad6" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.827305 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6jlnq"] Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.839912 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6jlnq"] Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.876094 4810 scope.go:117] "RemoveContainer" containerID="f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c" Sep 30 08:47:49 crc kubenswrapper[4810]: E0930 08:47:49.876681 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c\": container with ID starting with f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c not found: ID does not exist" containerID="f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.876718 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c"} err="failed to get container status \"f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c\": rpc error: code = NotFound desc = could not find container \"f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c\": container with ID starting with f5bc99b6cc57184dcfe1a0e155cdf8dc2ccae3a08b8663e83fa6809ad872545c not found: ID does not exist" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.876744 4810 scope.go:117] "RemoveContainer" containerID="8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513" Sep 30 08:47:49 crc kubenswrapper[4810]: E0930 08:47:49.877485 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513\": container with ID starting with 8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513 not found: ID does not exist" containerID="8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.877517 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513"} err="failed to get container status \"8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513\": rpc error: code = NotFound desc = could not find container \"8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513\": container with ID starting with 8e86abb3b93601bd33b821638f86ef52d5fc4a689ff0e9c0141c1a1796fb1513 not found: ID does not exist" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.877533 4810 scope.go:117] "RemoveContainer" containerID="2231513c8b90134e07c02b2e618fb90c10c3f43d14cf852985d5ab9a8ffa7ad6" Sep 30 08:47:49 crc kubenswrapper[4810]: E0930 08:47:49.877758 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2231513c8b90134e07c02b2e618fb90c10c3f43d14cf852985d5ab9a8ffa7ad6\": container with ID starting with 2231513c8b90134e07c02b2e618fb90c10c3f43d14cf852985d5ab9a8ffa7ad6 not found: ID does not exist" containerID="2231513c8b90134e07c02b2e618fb90c10c3f43d14cf852985d5ab9a8ffa7ad6" Sep 30 08:47:49 crc kubenswrapper[4810]: I0930 08:47:49.877786 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2231513c8b90134e07c02b2e618fb90c10c3f43d14cf852985d5ab9a8ffa7ad6"} err="failed to get container status \"2231513c8b90134e07c02b2e618fb90c10c3f43d14cf852985d5ab9a8ffa7ad6\": rpc error: code = NotFound desc = could not find container \"2231513c8b90134e07c02b2e618fb90c10c3f43d14cf852985d5ab9a8ffa7ad6\": container with ID starting with 2231513c8b90134e07c02b2e618fb90c10c3f43d14cf852985d5ab9a8ffa7ad6 not found: ID does not exist" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.624036 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k6m7g"] Sep 30 08:47:50 crc kubenswrapper[4810]: E0930 08:47:50.624895 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" containerName="registry-server" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.624927 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" containerName="registry-server" Sep 30 08:47:50 crc kubenswrapper[4810]: E0930 08:47:50.624954 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" containerName="extract-utilities" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.624968 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" containerName="extract-utilities" Sep 30 08:47:50 crc kubenswrapper[4810]: E0930 08:47:50.624995 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" containerName="extract-content" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.625006 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" containerName="extract-content" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.625351 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" containerName="registry-server" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.627421 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.642459 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k6m7g"] Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.664685 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-utilities\") pod \"redhat-operators-k6m7g\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.664788 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xvh2\" (UniqueName: \"kubernetes.io/projected/fa6eca49-c738-41e6-bfda-134db27352bc-kube-api-access-4xvh2\") pod \"redhat-operators-k6m7g\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.664909 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-catalog-content\") pod \"redhat-operators-k6m7g\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.766575 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-utilities\") pod \"redhat-operators-k6m7g\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.766653 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xvh2\" (UniqueName: \"kubernetes.io/projected/fa6eca49-c738-41e6-bfda-134db27352bc-kube-api-access-4xvh2\") pod \"redhat-operators-k6m7g\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.766703 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-catalog-content\") pod \"redhat-operators-k6m7g\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.767135 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-utilities\") pod \"redhat-operators-k6m7g\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.767167 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-catalog-content\") pod \"redhat-operators-k6m7g\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.788795 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xvh2\" (UniqueName: \"kubernetes.io/projected/fa6eca49-c738-41e6-bfda-134db27352bc-kube-api-access-4xvh2\") pod \"redhat-operators-k6m7g\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:47:50 crc kubenswrapper[4810]: I0930 08:47:50.998032 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:47:51 crc kubenswrapper[4810]: I0930 08:47:51.320135 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c" path="/var/lib/kubelet/pods/cb2cf6a7-b54e-4819-8b1d-f2a89bc7d77c/volumes" Sep 30 08:47:51 crc kubenswrapper[4810]: I0930 08:47:51.443919 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k6m7g"] Sep 30 08:47:51 crc kubenswrapper[4810]: I0930 08:47:51.772182 4810 generic.go:334] "Generic (PLEG): container finished" podID="fa6eca49-c738-41e6-bfda-134db27352bc" containerID="964f43b6b2db911d9d9c9e00446f1072f098b352ac9e90a281d16401f1e94703" exitCode=0 Sep 30 08:47:51 crc kubenswrapper[4810]: I0930 08:47:51.772239 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6m7g" event={"ID":"fa6eca49-c738-41e6-bfda-134db27352bc","Type":"ContainerDied","Data":"964f43b6b2db911d9d9c9e00446f1072f098b352ac9e90a281d16401f1e94703"} Sep 30 08:47:51 crc kubenswrapper[4810]: I0930 08:47:51.772518 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6m7g" event={"ID":"fa6eca49-c738-41e6-bfda-134db27352bc","Type":"ContainerStarted","Data":"2657acd00503727a1239796c2616b378995bbcf6d84605c85221172dd45ce18a"} Sep 30 08:47:56 crc kubenswrapper[4810]: I0930 08:47:56.306864 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:47:56 crc kubenswrapper[4810]: E0930 08:47:56.307702 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:48:03 crc kubenswrapper[4810]: I0930 08:48:03.906021 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6m7g" event={"ID":"fa6eca49-c738-41e6-bfda-134db27352bc","Type":"ContainerStarted","Data":"52ef7609e74159d0646144324670786db8077f570001a01270879c9aed1b2b3b"} Sep 30 08:48:05 crc kubenswrapper[4810]: I0930 08:48:05.934104 4810 generic.go:334] "Generic (PLEG): container finished" podID="fa6eca49-c738-41e6-bfda-134db27352bc" containerID="52ef7609e74159d0646144324670786db8077f570001a01270879c9aed1b2b3b" exitCode=0 Sep 30 08:48:05 crc kubenswrapper[4810]: I0930 08:48:05.934228 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6m7g" event={"ID":"fa6eca49-c738-41e6-bfda-134db27352bc","Type":"ContainerDied","Data":"52ef7609e74159d0646144324670786db8077f570001a01270879c9aed1b2b3b"} Sep 30 08:48:06 crc kubenswrapper[4810]: I0930 08:48:06.945696 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6m7g" event={"ID":"fa6eca49-c738-41e6-bfda-134db27352bc","Type":"ContainerStarted","Data":"e8b89276c8b32befe99256a44f1054ef92c71828bc922ddb2fa52bcaf99fb577"} Sep 30 08:48:06 crc kubenswrapper[4810]: I0930 08:48:06.981720 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k6m7g" podStartSLOduration=2.207970046 podStartE2EDuration="16.981692571s" podCreationTimestamp="2025-09-30 08:47:50 +0000 UTC" firstStartedPulling="2025-09-30 08:47:51.774059175 +0000 UTC m=+2695.226258452" lastFinishedPulling="2025-09-30 08:48:06.54778168 +0000 UTC m=+2709.999980977" observedRunningTime="2025-09-30 08:48:06.974129912 +0000 UTC m=+2710.426329209" watchObservedRunningTime="2025-09-30 08:48:06.981692571 +0000 UTC m=+2710.433891848" Sep 30 08:48:10 crc kubenswrapper[4810]: I0930 08:48:10.998326 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:48:10 crc kubenswrapper[4810]: I0930 08:48:10.998890 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:48:11 crc kubenswrapper[4810]: I0930 08:48:11.307196 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:48:11 crc kubenswrapper[4810]: E0930 08:48:11.307718 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:48:12 crc kubenswrapper[4810]: I0930 08:48:12.069580 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k6m7g" podUID="fa6eca49-c738-41e6-bfda-134db27352bc" containerName="registry-server" probeResult="failure" output=< Sep 30 08:48:12 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Sep 30 08:48:12 crc kubenswrapper[4810]: > Sep 30 08:48:21 crc kubenswrapper[4810]: I0930 08:48:21.078510 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:48:21 crc kubenswrapper[4810]: I0930 08:48:21.175041 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 08:48:21 crc kubenswrapper[4810]: I0930 08:48:21.650609 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k6m7g"] Sep 30 08:48:21 crc kubenswrapper[4810]: I0930 08:48:21.852300 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n6q2z"] Sep 30 08:48:21 crc kubenswrapper[4810]: I0930 08:48:21.852833 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n6q2z" podUID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerName="registry-server" containerID="cri-o://b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c" gracePeriod=2 Sep 30 08:48:22 crc kubenswrapper[4810]: E0930 08:48:22.019476 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c is running failed: container process not found" containerID="b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c" cmd=["grpc_health_probe","-addr=:50051"] Sep 30 08:48:22 crc kubenswrapper[4810]: E0930 08:48:22.020035 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c is running failed: container process not found" containerID="b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c" cmd=["grpc_health_probe","-addr=:50051"] Sep 30 08:48:22 crc kubenswrapper[4810]: E0930 08:48:22.020500 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c is running failed: container process not found" containerID="b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c" cmd=["grpc_health_probe","-addr=:50051"] Sep 30 08:48:22 crc kubenswrapper[4810]: E0930 08:48:22.020571 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-n6q2z" podUID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerName="registry-server" Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.141951 4810 generic.go:334] "Generic (PLEG): container finished" podID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerID="b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c" exitCode=0 Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.141998 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6q2z" event={"ID":"dad7c717-3a00-42d8-b71c-d2fb42083e83","Type":"ContainerDied","Data":"b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c"} Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.318118 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.468371 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fknvd\" (UniqueName: \"kubernetes.io/projected/dad7c717-3a00-42d8-b71c-d2fb42083e83-kube-api-access-fknvd\") pod \"dad7c717-3a00-42d8-b71c-d2fb42083e83\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.468622 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-utilities\") pod \"dad7c717-3a00-42d8-b71c-d2fb42083e83\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.468705 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-catalog-content\") pod \"dad7c717-3a00-42d8-b71c-d2fb42083e83\" (UID: \"dad7c717-3a00-42d8-b71c-d2fb42083e83\") " Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.470103 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-utilities" (OuterVolumeSpecName: "utilities") pod "dad7c717-3a00-42d8-b71c-d2fb42083e83" (UID: "dad7c717-3a00-42d8-b71c-d2fb42083e83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.475795 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dad7c717-3a00-42d8-b71c-d2fb42083e83-kube-api-access-fknvd" (OuterVolumeSpecName: "kube-api-access-fknvd") pod "dad7c717-3a00-42d8-b71c-d2fb42083e83" (UID: "dad7c717-3a00-42d8-b71c-d2fb42083e83"). InnerVolumeSpecName "kube-api-access-fknvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.570809 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fknvd\" (UniqueName: \"kubernetes.io/projected/dad7c717-3a00-42d8-b71c-d2fb42083e83-kube-api-access-fknvd\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.570838 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.575896 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dad7c717-3a00-42d8-b71c-d2fb42083e83" (UID: "dad7c717-3a00-42d8-b71c-d2fb42083e83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:48:22 crc kubenswrapper[4810]: I0930 08:48:22.672933 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dad7c717-3a00-42d8-b71c-d2fb42083e83-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:23 crc kubenswrapper[4810]: I0930 08:48:23.155009 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6q2z" event={"ID":"dad7c717-3a00-42d8-b71c-d2fb42083e83","Type":"ContainerDied","Data":"1f8bad1017a85035c6ed1696eba78a22a2f64e65cad1feecd243bd28c4ba880f"} Sep 30 08:48:23 crc kubenswrapper[4810]: I0930 08:48:23.156199 4810 scope.go:117] "RemoveContainer" containerID="b750a76e130950bcb93b1f35d88694603332ce379d1602572a6700bb5f5b152c" Sep 30 08:48:23 crc kubenswrapper[4810]: I0930 08:48:23.155387 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6q2z" Sep 30 08:48:23 crc kubenswrapper[4810]: I0930 08:48:23.198346 4810 scope.go:117] "RemoveContainer" containerID="b44c192827634d889a6685c8a4a8cca5a58be2984c28befb03c4924e9f1921db" Sep 30 08:48:23 crc kubenswrapper[4810]: I0930 08:48:23.203611 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n6q2z"] Sep 30 08:48:23 crc kubenswrapper[4810]: I0930 08:48:23.212404 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n6q2z"] Sep 30 08:48:23 crc kubenswrapper[4810]: I0930 08:48:23.226094 4810 scope.go:117] "RemoveContainer" containerID="2fbd121e378b38d23500b5250a8b553ae9c55b56bdae973cf1503c04e10b513d" Sep 30 08:48:23 crc kubenswrapper[4810]: I0930 08:48:23.323054 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dad7c717-3a00-42d8-b71c-d2fb42083e83" path="/var/lib/kubelet/pods/dad7c717-3a00-42d8-b71c-d2fb42083e83/volumes" Sep 30 08:48:24 crc kubenswrapper[4810]: I0930 08:48:24.306696 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:48:25 crc kubenswrapper[4810]: I0930 08:48:25.181578 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"2eb7a078098a0ac0ba80efe6dd4de079b7067414364ed78b82b35d53152b34ea"} Sep 30 08:48:39 crc kubenswrapper[4810]: I0930 08:48:39.363193 4810 generic.go:334] "Generic (PLEG): container finished" podID="8a25c107-d531-4a79-97e9-7201d6a8485a" containerID="c91606951ae9cf90fb0c34347a580a07530f4b5b2062458a0481adcdcbe57e18" exitCode=0 Sep 30 08:48:39 crc kubenswrapper[4810]: I0930 08:48:39.363322 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" event={"ID":"8a25c107-d531-4a79-97e9-7201d6a8485a","Type":"ContainerDied","Data":"c91606951ae9cf90fb0c34347a580a07530f4b5b2062458a0481adcdcbe57e18"} Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.805996 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.904815 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-0\") pod \"8a25c107-d531-4a79-97e9-7201d6a8485a\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.905443 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-1\") pod \"8a25c107-d531-4a79-97e9-7201d6a8485a\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.905579 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-extra-config-0\") pod \"8a25c107-d531-4a79-97e9-7201d6a8485a\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.905649 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-inventory\") pod \"8a25c107-d531-4a79-97e9-7201d6a8485a\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.905716 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdhpl\" (UniqueName: \"kubernetes.io/projected/8a25c107-d531-4a79-97e9-7201d6a8485a-kube-api-access-pdhpl\") pod \"8a25c107-d531-4a79-97e9-7201d6a8485a\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.905930 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-1\") pod \"8a25c107-d531-4a79-97e9-7201d6a8485a\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.906403 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-combined-ca-bundle\") pod \"8a25c107-d531-4a79-97e9-7201d6a8485a\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.906446 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-ssh-key\") pod \"8a25c107-d531-4a79-97e9-7201d6a8485a\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.906500 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-0\") pod \"8a25c107-d531-4a79-97e9-7201d6a8485a\" (UID: \"8a25c107-d531-4a79-97e9-7201d6a8485a\") " Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.913081 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a25c107-d531-4a79-97e9-7201d6a8485a-kube-api-access-pdhpl" (OuterVolumeSpecName: "kube-api-access-pdhpl") pod "8a25c107-d531-4a79-97e9-7201d6a8485a" (UID: "8a25c107-d531-4a79-97e9-7201d6a8485a"). InnerVolumeSpecName "kube-api-access-pdhpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.914669 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "8a25c107-d531-4a79-97e9-7201d6a8485a" (UID: "8a25c107-d531-4a79-97e9-7201d6a8485a"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.941625 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "8a25c107-d531-4a79-97e9-7201d6a8485a" (UID: "8a25c107-d531-4a79-97e9-7201d6a8485a"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.946513 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "8a25c107-d531-4a79-97e9-7201d6a8485a" (UID: "8a25c107-d531-4a79-97e9-7201d6a8485a"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.952109 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "8a25c107-d531-4a79-97e9-7201d6a8485a" (UID: "8a25c107-d531-4a79-97e9-7201d6a8485a"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.954890 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-inventory" (OuterVolumeSpecName: "inventory") pod "8a25c107-d531-4a79-97e9-7201d6a8485a" (UID: "8a25c107-d531-4a79-97e9-7201d6a8485a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.956525 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "8a25c107-d531-4a79-97e9-7201d6a8485a" (UID: "8a25c107-d531-4a79-97e9-7201d6a8485a"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.962076 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "8a25c107-d531-4a79-97e9-7201d6a8485a" (UID: "8a25c107-d531-4a79-97e9-7201d6a8485a"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:48:40 crc kubenswrapper[4810]: I0930 08:48:40.971657 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8a25c107-d531-4a79-97e9-7201d6a8485a" (UID: "8a25c107-d531-4a79-97e9-7201d6a8485a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.014173 4810 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.014245 4810 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.014303 4810 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.014332 4810 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.014364 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.014391 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdhpl\" (UniqueName: \"kubernetes.io/projected/8a25c107-d531-4a79-97e9-7201d6a8485a-kube-api-access-pdhpl\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.014421 4810 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.014447 4810 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.014472 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a25c107-d531-4a79-97e9-7201d6a8485a-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.390217 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" event={"ID":"8a25c107-d531-4a79-97e9-7201d6a8485a","Type":"ContainerDied","Data":"4033261d3c6d74bee294b2cd87b9613c94992498d1cb2b09a8dc556e9efb5cd2"} Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.390255 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4033261d3c6d74bee294b2cd87b9613c94992498d1cb2b09a8dc556e9efb5cd2" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.390340 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v6q68" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.547994 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r"] Sep 30 08:48:41 crc kubenswrapper[4810]: E0930 08:48:41.548537 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerName="extract-utilities" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.548564 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerName="extract-utilities" Sep 30 08:48:41 crc kubenswrapper[4810]: E0930 08:48:41.548588 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a25c107-d531-4a79-97e9-7201d6a8485a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.548596 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a25c107-d531-4a79-97e9-7201d6a8485a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 08:48:41 crc kubenswrapper[4810]: E0930 08:48:41.548622 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerName="extract-content" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.548631 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerName="extract-content" Sep 30 08:48:41 crc kubenswrapper[4810]: E0930 08:48:41.548651 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerName="registry-server" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.548658 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerName="registry-server" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.548942 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="dad7c717-3a00-42d8-b71c-d2fb42083e83" containerName="registry-server" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.548964 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a25c107-d531-4a79-97e9-7201d6a8485a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.549899 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.564618 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.564625 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.565452 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.566393 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.566398 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-svc7t" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.581017 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r"] Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.625243 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.625295 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8bpm\" (UniqueName: \"kubernetes.io/projected/ba3931a9-05a1-47f8-a7f7-f41a477a164b-kube-api-access-j8bpm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.625344 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.625388 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.625406 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.625427 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.625522 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.727365 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.727407 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8bpm\" (UniqueName: \"kubernetes.io/projected/ba3931a9-05a1-47f8-a7f7-f41a477a164b-kube-api-access-j8bpm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.727458 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.727502 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.727521 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.727543 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.727601 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.731754 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.731881 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.732473 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.733553 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.744754 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.745384 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.748983 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8bpm\" (UniqueName: \"kubernetes.io/projected/ba3931a9-05a1-47f8-a7f7-f41a477a164b-kube-api-access-j8bpm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:41 crc kubenswrapper[4810]: I0930 08:48:41.874929 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:48:42 crc kubenswrapper[4810]: I0930 08:48:42.279851 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r"] Sep 30 08:48:42 crc kubenswrapper[4810]: I0930 08:48:42.400366 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" event={"ID":"ba3931a9-05a1-47f8-a7f7-f41a477a164b","Type":"ContainerStarted","Data":"038543056f159a66013db103ffa024006c3e22d7541477eb743d11ad92d74a4f"} Sep 30 08:48:43 crc kubenswrapper[4810]: I0930 08:48:43.411312 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" event={"ID":"ba3931a9-05a1-47f8-a7f7-f41a477a164b","Type":"ContainerStarted","Data":"4cb61298bcd5c0b9087ee98db4cd027f8fe41ae55c7f16d22c320177c920d27e"} Sep 30 08:48:43 crc kubenswrapper[4810]: I0930 08:48:43.436891 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" podStartSLOduration=1.7326454359999999 podStartE2EDuration="2.436870017s" podCreationTimestamp="2025-09-30 08:48:41 +0000 UTC" firstStartedPulling="2025-09-30 08:48:42.279715222 +0000 UTC m=+2745.731914509" lastFinishedPulling="2025-09-30 08:48:42.983939813 +0000 UTC m=+2746.436139090" observedRunningTime="2025-09-30 08:48:43.43216345 +0000 UTC m=+2746.884362727" watchObservedRunningTime="2025-09-30 08:48:43.436870017 +0000 UTC m=+2746.889069294" Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.471511 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zl5x9"] Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.475612 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.518248 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zl5x9"] Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.556116 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7adac989-68a1-4f65-b249-7f43d8bac8be-utilities\") pod \"community-operators-zl5x9\" (UID: \"7adac989-68a1-4f65-b249-7f43d8bac8be\") " pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.556301 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6mxm\" (UniqueName: \"kubernetes.io/projected/7adac989-68a1-4f65-b249-7f43d8bac8be-kube-api-access-j6mxm\") pod \"community-operators-zl5x9\" (UID: \"7adac989-68a1-4f65-b249-7f43d8bac8be\") " pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.556341 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7adac989-68a1-4f65-b249-7f43d8bac8be-catalog-content\") pod \"community-operators-zl5x9\" (UID: \"7adac989-68a1-4f65-b249-7f43d8bac8be\") " pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.658442 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6mxm\" (UniqueName: \"kubernetes.io/projected/7adac989-68a1-4f65-b249-7f43d8bac8be-kube-api-access-j6mxm\") pod \"community-operators-zl5x9\" (UID: \"7adac989-68a1-4f65-b249-7f43d8bac8be\") " pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.658514 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7adac989-68a1-4f65-b249-7f43d8bac8be-catalog-content\") pod \"community-operators-zl5x9\" (UID: \"7adac989-68a1-4f65-b249-7f43d8bac8be\") " pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.658661 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7adac989-68a1-4f65-b249-7f43d8bac8be-utilities\") pod \"community-operators-zl5x9\" (UID: \"7adac989-68a1-4f65-b249-7f43d8bac8be\") " pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.659209 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7adac989-68a1-4f65-b249-7f43d8bac8be-catalog-content\") pod \"community-operators-zl5x9\" (UID: \"7adac989-68a1-4f65-b249-7f43d8bac8be\") " pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.659353 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7adac989-68a1-4f65-b249-7f43d8bac8be-utilities\") pod \"community-operators-zl5x9\" (UID: \"7adac989-68a1-4f65-b249-7f43d8bac8be\") " pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.678467 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6mxm\" (UniqueName: \"kubernetes.io/projected/7adac989-68a1-4f65-b249-7f43d8bac8be-kube-api-access-j6mxm\") pod \"community-operators-zl5x9\" (UID: \"7adac989-68a1-4f65-b249-7f43d8bac8be\") " pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:22 crc kubenswrapper[4810]: I0930 08:49:22.807098 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:23 crc kubenswrapper[4810]: I0930 08:49:23.399835 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zl5x9"] Sep 30 08:49:23 crc kubenswrapper[4810]: I0930 08:49:23.902178 4810 generic.go:334] "Generic (PLEG): container finished" podID="7adac989-68a1-4f65-b249-7f43d8bac8be" containerID="351de1db29eb2400b20638ab6cd238648aa24ca6608542203976b3946f79650e" exitCode=0 Sep 30 08:49:23 crc kubenswrapper[4810]: I0930 08:49:23.902308 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zl5x9" event={"ID":"7adac989-68a1-4f65-b249-7f43d8bac8be","Type":"ContainerDied","Data":"351de1db29eb2400b20638ab6cd238648aa24ca6608542203976b3946f79650e"} Sep 30 08:49:23 crc kubenswrapper[4810]: I0930 08:49:23.902515 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zl5x9" event={"ID":"7adac989-68a1-4f65-b249-7f43d8bac8be","Type":"ContainerStarted","Data":"606b01d6c12473b9f9e8644a347cf7d565d581e476bf54790955d20d81e71ff3"} Sep 30 08:49:28 crc kubenswrapper[4810]: I0930 08:49:28.957331 4810 generic.go:334] "Generic (PLEG): container finished" podID="7adac989-68a1-4f65-b249-7f43d8bac8be" containerID="f161b2a94f0aa7ea14b3c7813aa5cf20cbc3a47431410749eecf4bdb7f49b96d" exitCode=0 Sep 30 08:49:28 crc kubenswrapper[4810]: I0930 08:49:28.957427 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zl5x9" event={"ID":"7adac989-68a1-4f65-b249-7f43d8bac8be","Type":"ContainerDied","Data":"f161b2a94f0aa7ea14b3c7813aa5cf20cbc3a47431410749eecf4bdb7f49b96d"} Sep 30 08:49:29 crc kubenswrapper[4810]: I0930 08:49:29.976338 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zl5x9" event={"ID":"7adac989-68a1-4f65-b249-7f43d8bac8be","Type":"ContainerStarted","Data":"ab39e3735bab46fb6349c316c67ceee90bb1dbe33beef58bbec4431dc221b0f6"} Sep 30 08:49:29 crc kubenswrapper[4810]: I0930 08:49:29.998922 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zl5x9" podStartSLOduration=2.534456372 podStartE2EDuration="7.998898947s" podCreationTimestamp="2025-09-30 08:49:22 +0000 UTC" firstStartedPulling="2025-09-30 08:49:23.905308096 +0000 UTC m=+2787.357507403" lastFinishedPulling="2025-09-30 08:49:29.369750701 +0000 UTC m=+2792.821949978" observedRunningTime="2025-09-30 08:49:29.997672742 +0000 UTC m=+2793.449872049" watchObservedRunningTime="2025-09-30 08:49:29.998898947 +0000 UTC m=+2793.451098234" Sep 30 08:49:32 crc kubenswrapper[4810]: I0930 08:49:32.807594 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:32 crc kubenswrapper[4810]: I0930 08:49:32.807897 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:32 crc kubenswrapper[4810]: I0930 08:49:32.866622 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:42 crc kubenswrapper[4810]: I0930 08:49:42.886774 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zl5x9" Sep 30 08:49:42 crc kubenswrapper[4810]: I0930 08:49:42.954501 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zl5x9"] Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.009478 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2lx8v"] Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.010058 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2lx8v" podUID="6596bb1a-28f7-453f-8b30-ade04d3180b0" containerName="registry-server" containerID="cri-o://57d19542556ac781b42a39a2c56d445960c973887d372238c996b162c2889992" gracePeriod=2 Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.168127 4810 generic.go:334] "Generic (PLEG): container finished" podID="6596bb1a-28f7-453f-8b30-ade04d3180b0" containerID="57d19542556ac781b42a39a2c56d445960c973887d372238c996b162c2889992" exitCode=0 Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.169161 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lx8v" event={"ID":"6596bb1a-28f7-453f-8b30-ade04d3180b0","Type":"ContainerDied","Data":"57d19542556ac781b42a39a2c56d445960c973887d372238c996b162c2889992"} Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.490231 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.570904 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-utilities\") pod \"6596bb1a-28f7-453f-8b30-ade04d3180b0\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.570956 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbqd8\" (UniqueName: \"kubernetes.io/projected/6596bb1a-28f7-453f-8b30-ade04d3180b0-kube-api-access-zbqd8\") pod \"6596bb1a-28f7-453f-8b30-ade04d3180b0\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.570986 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-catalog-content\") pod \"6596bb1a-28f7-453f-8b30-ade04d3180b0\" (UID: \"6596bb1a-28f7-453f-8b30-ade04d3180b0\") " Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.571696 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-utilities" (OuterVolumeSpecName: "utilities") pod "6596bb1a-28f7-453f-8b30-ade04d3180b0" (UID: "6596bb1a-28f7-453f-8b30-ade04d3180b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.577153 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6596bb1a-28f7-453f-8b30-ade04d3180b0-kube-api-access-zbqd8" (OuterVolumeSpecName: "kube-api-access-zbqd8") pod "6596bb1a-28f7-453f-8b30-ade04d3180b0" (UID: "6596bb1a-28f7-453f-8b30-ade04d3180b0"). InnerVolumeSpecName "kube-api-access-zbqd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.622375 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6596bb1a-28f7-453f-8b30-ade04d3180b0" (UID: "6596bb1a-28f7-453f-8b30-ade04d3180b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.673468 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.673761 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbqd8\" (UniqueName: \"kubernetes.io/projected/6596bb1a-28f7-453f-8b30-ade04d3180b0-kube-api-access-zbqd8\") on node \"crc\" DevicePath \"\"" Sep 30 08:49:43 crc kubenswrapper[4810]: I0930 08:49:43.673843 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6596bb1a-28f7-453f-8b30-ade04d3180b0-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:49:44 crc kubenswrapper[4810]: I0930 08:49:44.182855 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lx8v" event={"ID":"6596bb1a-28f7-453f-8b30-ade04d3180b0","Type":"ContainerDied","Data":"81c4f8c2b311128c7643e911bd8f85d80cfe407b56fea94d5ffe6d305ade56d2"} Sep 30 08:49:44 crc kubenswrapper[4810]: I0930 08:49:44.182908 4810 scope.go:117] "RemoveContainer" containerID="57d19542556ac781b42a39a2c56d445960c973887d372238c996b162c2889992" Sep 30 08:49:44 crc kubenswrapper[4810]: I0930 08:49:44.184419 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lx8v" Sep 30 08:49:44 crc kubenswrapper[4810]: I0930 08:49:44.218744 4810 scope.go:117] "RemoveContainer" containerID="4127fa66be8be0eca71c51c0a3d2849ecb0f83fa49390811d02f2d3cc19e8f1f" Sep 30 08:49:44 crc kubenswrapper[4810]: I0930 08:49:44.231995 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2lx8v"] Sep 30 08:49:44 crc kubenswrapper[4810]: I0930 08:49:44.246569 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2lx8v"] Sep 30 08:49:44 crc kubenswrapper[4810]: I0930 08:49:44.266439 4810 scope.go:117] "RemoveContainer" containerID="01c1c7a0b594e157177ef4fa261ee3ab0d78cf0c7f6071e18b59fa53e641000e" Sep 30 08:49:45 crc kubenswrapper[4810]: I0930 08:49:45.319663 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6596bb1a-28f7-453f-8b30-ade04d3180b0" path="/var/lib/kubelet/pods/6596bb1a-28f7-453f-8b30-ade04d3180b0/volumes" Sep 30 08:50:45 crc kubenswrapper[4810]: I0930 08:50:45.911237 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:50:45 crc kubenswrapper[4810]: I0930 08:50:45.911952 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:51:15 crc kubenswrapper[4810]: I0930 08:51:15.911774 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:51:15 crc kubenswrapper[4810]: I0930 08:51:15.912322 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:51:20 crc kubenswrapper[4810]: I0930 08:51:20.366602 4810 generic.go:334] "Generic (PLEG): container finished" podID="ba3931a9-05a1-47f8-a7f7-f41a477a164b" containerID="4cb61298bcd5c0b9087ee98db4cd027f8fe41ae55c7f16d22c320177c920d27e" exitCode=0 Sep 30 08:51:20 crc kubenswrapper[4810]: I0930 08:51:20.366680 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" event={"ID":"ba3931a9-05a1-47f8-a7f7-f41a477a164b","Type":"ContainerDied","Data":"4cb61298bcd5c0b9087ee98db4cd027f8fe41ae55c7f16d22c320177c920d27e"} Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.887095 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.941097 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-0\") pod \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.941190 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-telemetry-combined-ca-bundle\") pod \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.941259 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-1\") pod \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.941319 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8bpm\" (UniqueName: \"kubernetes.io/projected/ba3931a9-05a1-47f8-a7f7-f41a477a164b-kube-api-access-j8bpm\") pod \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.941377 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-2\") pod \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.941396 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-inventory\") pod \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.941443 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ssh-key\") pod \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\" (UID: \"ba3931a9-05a1-47f8-a7f7-f41a477a164b\") " Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.949897 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba3931a9-05a1-47f8-a7f7-f41a477a164b-kube-api-access-j8bpm" (OuterVolumeSpecName: "kube-api-access-j8bpm") pod "ba3931a9-05a1-47f8-a7f7-f41a477a164b" (UID: "ba3931a9-05a1-47f8-a7f7-f41a477a164b"). InnerVolumeSpecName "kube-api-access-j8bpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.954549 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "ba3931a9-05a1-47f8-a7f7-f41a477a164b" (UID: "ba3931a9-05a1-47f8-a7f7-f41a477a164b"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.976489 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ba3931a9-05a1-47f8-a7f7-f41a477a164b" (UID: "ba3931a9-05a1-47f8-a7f7-f41a477a164b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.976552 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "ba3931a9-05a1-47f8-a7f7-f41a477a164b" (UID: "ba3931a9-05a1-47f8-a7f7-f41a477a164b"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.979452 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-inventory" (OuterVolumeSpecName: "inventory") pod "ba3931a9-05a1-47f8-a7f7-f41a477a164b" (UID: "ba3931a9-05a1-47f8-a7f7-f41a477a164b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.991055 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "ba3931a9-05a1-47f8-a7f7-f41a477a164b" (UID: "ba3931a9-05a1-47f8-a7f7-f41a477a164b"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:51:21 crc kubenswrapper[4810]: I0930 08:51:21.992607 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "ba3931a9-05a1-47f8-a7f7-f41a477a164b" (UID: "ba3931a9-05a1-47f8-a7f7-f41a477a164b"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:51:22 crc kubenswrapper[4810]: I0930 08:51:22.044547 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 08:51:22 crc kubenswrapper[4810]: I0930 08:51:22.044603 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:51:22 crc kubenswrapper[4810]: I0930 08:51:22.044626 4810 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:51:22 crc kubenswrapper[4810]: I0930 08:51:22.044654 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Sep 30 08:51:22 crc kubenswrapper[4810]: I0930 08:51:22.044673 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8bpm\" (UniqueName: \"kubernetes.io/projected/ba3931a9-05a1-47f8-a7f7-f41a477a164b-kube-api-access-j8bpm\") on node \"crc\" DevicePath \"\"" Sep 30 08:51:22 crc kubenswrapper[4810]: I0930 08:51:22.044692 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Sep 30 08:51:22 crc kubenswrapper[4810]: I0930 08:51:22.044710 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba3931a9-05a1-47f8-a7f7-f41a477a164b-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 08:51:22 crc kubenswrapper[4810]: I0930 08:51:22.397973 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" event={"ID":"ba3931a9-05a1-47f8-a7f7-f41a477a164b","Type":"ContainerDied","Data":"038543056f159a66013db103ffa024006c3e22d7541477eb743d11ad92d74a4f"} Sep 30 08:51:22 crc kubenswrapper[4810]: I0930 08:51:22.398041 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="038543056f159a66013db103ffa024006c3e22d7541477eb743d11ad92d74a4f" Sep 30 08:51:22 crc kubenswrapper[4810]: I0930 08:51:22.398050 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r" Sep 30 08:51:45 crc kubenswrapper[4810]: I0930 08:51:45.912138 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:51:45 crc kubenswrapper[4810]: I0930 08:51:45.912792 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:51:45 crc kubenswrapper[4810]: I0930 08:51:45.912865 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:51:45 crc kubenswrapper[4810]: I0930 08:51:45.913895 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2eb7a078098a0ac0ba80efe6dd4de079b7067414364ed78b82b35d53152b34ea"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:51:45 crc kubenswrapper[4810]: I0930 08:51:45.914002 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://2eb7a078098a0ac0ba80efe6dd4de079b7067414364ed78b82b35d53152b34ea" gracePeriod=600 Sep 30 08:51:46 crc kubenswrapper[4810]: I0930 08:51:46.686983 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="2eb7a078098a0ac0ba80efe6dd4de079b7067414364ed78b82b35d53152b34ea" exitCode=0 Sep 30 08:51:46 crc kubenswrapper[4810]: I0930 08:51:46.687074 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"2eb7a078098a0ac0ba80efe6dd4de079b7067414364ed78b82b35d53152b34ea"} Sep 30 08:51:46 crc kubenswrapper[4810]: I0930 08:51:46.687570 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5"} Sep 30 08:51:46 crc kubenswrapper[4810]: I0930 08:51:46.687599 4810 scope.go:117] "RemoveContainer" containerID="0202ddd88ba4bdebbdbf569bbe6c9dfbba06152f1da100789af777166b4cfa9d" Sep 30 08:51:58 crc kubenswrapper[4810]: I0930 08:51:58.833631 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:51:58 crc kubenswrapper[4810]: I0930 08:51:58.834288 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="prometheus" containerID="cri-o://66e23df6841865f3e52782074053a40ffa2afa2c03e76c69473fa0655c6d52f3" gracePeriod=600 Sep 30 08:51:58 crc kubenswrapper[4810]: I0930 08:51:58.834390 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="config-reloader" containerID="cri-o://a274e25cded7421e038e2251ec27af7dbe8172d4c7e4d2c06168aa669836f3d0" gracePeriod=600 Sep 30 08:51:58 crc kubenswrapper[4810]: I0930 08:51:58.834390 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="thanos-sidecar" containerID="cri-o://612a628dc5694fefa9476ba97ed57198bc80e617d9dbe0dcfb09e14d0b996d8d" gracePeriod=600 Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.843520 4810 generic.go:334] "Generic (PLEG): container finished" podID="eb9bba80-4046-4788-ac20-c682625132d8" containerID="612a628dc5694fefa9476ba97ed57198bc80e617d9dbe0dcfb09e14d0b996d8d" exitCode=0 Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.844143 4810 generic.go:334] "Generic (PLEG): container finished" podID="eb9bba80-4046-4788-ac20-c682625132d8" containerID="a274e25cded7421e038e2251ec27af7dbe8172d4c7e4d2c06168aa669836f3d0" exitCode=0 Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.844154 4810 generic.go:334] "Generic (PLEG): container finished" podID="eb9bba80-4046-4788-ac20-c682625132d8" containerID="66e23df6841865f3e52782074053a40ffa2afa2c03e76c69473fa0655c6d52f3" exitCode=0 Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.844102 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb9bba80-4046-4788-ac20-c682625132d8","Type":"ContainerDied","Data":"612a628dc5694fefa9476ba97ed57198bc80e617d9dbe0dcfb09e14d0b996d8d"} Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.844193 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb9bba80-4046-4788-ac20-c682625132d8","Type":"ContainerDied","Data":"a274e25cded7421e038e2251ec27af7dbe8172d4c7e4d2c06168aa669836f3d0"} Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.844207 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb9bba80-4046-4788-ac20-c682625132d8","Type":"ContainerDied","Data":"66e23df6841865f3e52782074053a40ffa2afa2c03e76c69473fa0655c6d52f3"} Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.844220 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb9bba80-4046-4788-ac20-c682625132d8","Type":"ContainerDied","Data":"adb297544aef1eeff8c20225e20a3e7b5541f56f960469d3e41d23e08b6ebc18"} Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.844232 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adb297544aef1eeff8c20225e20a3e7b5541f56f960469d3e41d23e08b6ebc18" Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.875286 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.903548 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eb9bba80-4046-4788-ac20-c682625132d8-prometheus-metric-storage-rulefiles-0\") pod \"eb9bba80-4046-4788-ac20-c682625132d8\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.903612 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eb9bba80-4046-4788-ac20-c682625132d8-config-out\") pod \"eb9bba80-4046-4788-ac20-c682625132d8\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.903680 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-tls-assets\") pod \"eb9bba80-4046-4788-ac20-c682625132d8\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.903709 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-config\") pod \"eb9bba80-4046-4788-ac20-c682625132d8\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.903737 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"eb9bba80-4046-4788-ac20-c682625132d8\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.903767 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config\") pod \"eb9bba80-4046-4788-ac20-c682625132d8\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.903898 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"eb9bba80-4046-4788-ac20-c682625132d8\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.903936 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24d77\" (UniqueName: \"kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-kube-api-access-24d77\") pod \"eb9bba80-4046-4788-ac20-c682625132d8\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.903972 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-secret-combined-ca-bundle\") pod \"eb9bba80-4046-4788-ac20-c682625132d8\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.903997 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-thanos-prometheus-http-client-file\") pod \"eb9bba80-4046-4788-ac20-c682625132d8\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.904039 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"eb9bba80-4046-4788-ac20-c682625132d8\" (UID: \"eb9bba80-4046-4788-ac20-c682625132d8\") " Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.904208 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb9bba80-4046-4788-ac20-c682625132d8-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "eb9bba80-4046-4788-ac20-c682625132d8" (UID: "eb9bba80-4046-4788-ac20-c682625132d8"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.904782 4810 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eb9bba80-4046-4788-ac20-c682625132d8-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.912979 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-kube-api-access-24d77" (OuterVolumeSpecName: "kube-api-access-24d77") pod "eb9bba80-4046-4788-ac20-c682625132d8" (UID: "eb9bba80-4046-4788-ac20-c682625132d8"). InnerVolumeSpecName "kube-api-access-24d77". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.918461 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "eb9bba80-4046-4788-ac20-c682625132d8" (UID: "eb9bba80-4046-4788-ac20-c682625132d8"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.918508 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "eb9bba80-4046-4788-ac20-c682625132d8" (UID: "eb9bba80-4046-4788-ac20-c682625132d8"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.918783 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb9bba80-4046-4788-ac20-c682625132d8-config-out" (OuterVolumeSpecName: "config-out") pod "eb9bba80-4046-4788-ac20-c682625132d8" (UID: "eb9bba80-4046-4788-ac20-c682625132d8"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.942417 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-config" (OuterVolumeSpecName: "config") pod "eb9bba80-4046-4788-ac20-c682625132d8" (UID: "eb9bba80-4046-4788-ac20-c682625132d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.975154 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "eb9bba80-4046-4788-ac20-c682625132d8" (UID: "eb9bba80-4046-4788-ac20-c682625132d8"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:51:59 crc kubenswrapper[4810]: I0930 08:51:59.975520 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "eb9bba80-4046-4788-ac20-c682625132d8" (UID: "eb9bba80-4046-4788-ac20-c682625132d8"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.008693 4810 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-tls-assets\") on node \"crc\" DevicePath \"\"" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.008717 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.008727 4810 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.008741 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24d77\" (UniqueName: \"kubernetes.io/projected/eb9bba80-4046-4788-ac20-c682625132d8-kube-api-access-24d77\") on node \"crc\" DevicePath \"\"" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.008752 4810 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.008762 4810 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.008789 4810 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eb9bba80-4046-4788-ac20-c682625132d8-config-out\") on node \"crc\" DevicePath \"\"" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.016117 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "eb9bba80-4046-4788-ac20-c682625132d8" (UID: "eb9bba80-4046-4788-ac20-c682625132d8"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.131163 4810 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.166707 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "eb9bba80-4046-4788-ac20-c682625132d8" (UID: "eb9bba80-4046-4788-ac20-c682625132d8"). InnerVolumeSpecName "pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.180034 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config" (OuterVolumeSpecName: "web-config") pod "eb9bba80-4046-4788-ac20-c682625132d8" (UID: "eb9bba80-4046-4788-ac20-c682625132d8"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.233315 4810 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eb9bba80-4046-4788-ac20-c682625132d8-web-config\") on node \"crc\" DevicePath \"\"" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.233369 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") on node \"crc\" " Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.262040 4810 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.263014 4810 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c") on node "crc" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.335948 4810 reconciler_common.go:293] "Volume detached for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") on node \"crc\" DevicePath \"\"" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.854684 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.902591 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.907765 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.936697 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:52:00 crc kubenswrapper[4810]: E0930 08:52:00.937332 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6596bb1a-28f7-453f-8b30-ade04d3180b0" containerName="extract-content" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.937402 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6596bb1a-28f7-453f-8b30-ade04d3180b0" containerName="extract-content" Sep 30 08:52:00 crc kubenswrapper[4810]: E0930 08:52:00.937473 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6596bb1a-28f7-453f-8b30-ade04d3180b0" containerName="extract-utilities" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.937523 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6596bb1a-28f7-453f-8b30-ade04d3180b0" containerName="extract-utilities" Sep 30 08:52:00 crc kubenswrapper[4810]: E0930 08:52:00.937571 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="config-reloader" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.937618 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="config-reloader" Sep 30 08:52:00 crc kubenswrapper[4810]: E0930 08:52:00.937692 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba3931a9-05a1-47f8-a7f7-f41a477a164b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.937743 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba3931a9-05a1-47f8-a7f7-f41a477a164b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 08:52:00 crc kubenswrapper[4810]: E0930 08:52:00.937819 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6596bb1a-28f7-453f-8b30-ade04d3180b0" containerName="registry-server" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.937883 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6596bb1a-28f7-453f-8b30-ade04d3180b0" containerName="registry-server" Sep 30 08:52:00 crc kubenswrapper[4810]: E0930 08:52:00.937939 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="prometheus" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.937988 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="prometheus" Sep 30 08:52:00 crc kubenswrapper[4810]: E0930 08:52:00.938044 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="init-config-reloader" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.938096 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="init-config-reloader" Sep 30 08:52:00 crc kubenswrapper[4810]: E0930 08:52:00.938153 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="thanos-sidecar" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.938202 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="thanos-sidecar" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.938458 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="thanos-sidecar" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.938536 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6596bb1a-28f7-453f-8b30-ade04d3180b0" containerName="registry-server" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.938597 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="config-reloader" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.938659 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba3931a9-05a1-47f8-a7f7-f41a477a164b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.938759 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9bba80-4046-4788-ac20-c682625132d8" containerName="prometheus" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.940576 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.944930 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.945165 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.945258 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.945367 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-t6zzk" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.945716 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Sep 30 08:52:00 crc kubenswrapper[4810]: I0930 08:52:00.962642 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.026573 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.049788 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.049844 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ad43140a-2b1c-443b-9595-901c71e14f0d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.050123 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ad43140a-2b1c-443b-9595-901c71e14f0d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.050876 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt4gz\" (UniqueName: \"kubernetes.io/projected/ad43140a-2b1c-443b-9595-901c71e14f0d-kube-api-access-mt4gz\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.052063 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ad43140a-2b1c-443b-9595-901c71e14f0d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.052117 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-config\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.053403 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.053475 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.053513 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.053645 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.053701 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.155935 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt4gz\" (UniqueName: \"kubernetes.io/projected/ad43140a-2b1c-443b-9595-901c71e14f0d-kube-api-access-mt4gz\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.155983 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ad43140a-2b1c-443b-9595-901c71e14f0d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.156005 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-config\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.156028 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.156050 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.156068 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.156093 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.156111 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.156172 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.156191 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ad43140a-2b1c-443b-9595-901c71e14f0d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.156218 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ad43140a-2b1c-443b-9595-901c71e14f0d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.158164 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ad43140a-2b1c-443b-9595-901c71e14f0d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.162522 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.162578 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b8ca484aa6b3c023d84489054d30b21e25eca5ebf5935b3c554e8125ba6153af/globalmount\"" pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.162990 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-config\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.164862 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.165208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.166514 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ad43140a-2b1c-443b-9595-901c71e14f0d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.168224 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ad43140a-2b1c-443b-9595-901c71e14f0d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.171697 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.171755 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.175681 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad43140a-2b1c-443b-9595-901c71e14f0d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.180342 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt4gz\" (UniqueName: \"kubernetes.io/projected/ad43140a-2b1c-443b-9595-901c71e14f0d-kube-api-access-mt4gz\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.228559 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2882f5ac-2eeb-4bb0-814e-4806ab0dae6c\") pod \"prometheus-metric-storage-0\" (UID: \"ad43140a-2b1c-443b-9595-901c71e14f0d\") " pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.317885 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.320487 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb9bba80-4046-4788-ac20-c682625132d8" path="/var/lib/kubelet/pods/eb9bba80-4046-4788-ac20-c682625132d8/volumes" Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.801531 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Sep 30 08:52:01 crc kubenswrapper[4810]: W0930 08:52:01.809885 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad43140a_2b1c_443b_9595_901c71e14f0d.slice/crio-a4a9765623a8547d400d5a608cca96a7ff34870213582d9dd123ada07bf6fc9b WatchSource:0}: Error finding container a4a9765623a8547d400d5a608cca96a7ff34870213582d9dd123ada07bf6fc9b: Status 404 returned error can't find the container with id a4a9765623a8547d400d5a608cca96a7ff34870213582d9dd123ada07bf6fc9b Sep 30 08:52:01 crc kubenswrapper[4810]: I0930 08:52:01.866555 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ad43140a-2b1c-443b-9595-901c71e14f0d","Type":"ContainerStarted","Data":"a4a9765623a8547d400d5a608cca96a7ff34870213582d9dd123ada07bf6fc9b"} Sep 30 08:52:05 crc kubenswrapper[4810]: I0930 08:52:05.915426 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ad43140a-2b1c-443b-9595-901c71e14f0d","Type":"ContainerStarted","Data":"693220d6b0647ff3783c000b2820e64b1ff3a3bd9f28276fb71d9679673a72c5"} Sep 30 08:52:17 crc kubenswrapper[4810]: I0930 08:52:17.043465 4810 generic.go:334] "Generic (PLEG): container finished" podID="ad43140a-2b1c-443b-9595-901c71e14f0d" containerID="693220d6b0647ff3783c000b2820e64b1ff3a3bd9f28276fb71d9679673a72c5" exitCode=0 Sep 30 08:52:17 crc kubenswrapper[4810]: I0930 08:52:17.043553 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ad43140a-2b1c-443b-9595-901c71e14f0d","Type":"ContainerDied","Data":"693220d6b0647ff3783c000b2820e64b1ff3a3bd9f28276fb71d9679673a72c5"} Sep 30 08:52:18 crc kubenswrapper[4810]: I0930 08:52:18.054639 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ad43140a-2b1c-443b-9595-901c71e14f0d","Type":"ContainerStarted","Data":"3a1f81d9a32e24579b525c3089a33b61cb4638ce446ed2e0de82d58adbf6cb1a"} Sep 30 08:52:21 crc kubenswrapper[4810]: I0930 08:52:21.098041 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ad43140a-2b1c-443b-9595-901c71e14f0d","Type":"ContainerStarted","Data":"ab1bd129c7bfbcf30584336949bb076c4882bce190dcf1670aa4b58398303f7d"} Sep 30 08:52:21 crc kubenswrapper[4810]: I0930 08:52:21.098536 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ad43140a-2b1c-443b-9595-901c71e14f0d","Type":"ContainerStarted","Data":"86fff87cded84ec512cbf93f7b1b5a422b32e12a4704f7355ea4a3ca25300238"} Sep 30 08:52:21 crc kubenswrapper[4810]: I0930 08:52:21.128492 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=21.128474472 podStartE2EDuration="21.128474472s" podCreationTimestamp="2025-09-30 08:52:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 08:52:21.127493214 +0000 UTC m=+2964.579692501" watchObservedRunningTime="2025-09-30 08:52:21.128474472 +0000 UTC m=+2964.580673739" Sep 30 08:52:21 crc kubenswrapper[4810]: I0930 08:52:21.325050 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:21 crc kubenswrapper[4810]: I0930 08:52:21.656377 4810 scope.go:117] "RemoveContainer" containerID="43a44738cda6d3139fcbce99404f6f250153b2205c24b1b2cda00b30fe0cb1ff" Sep 30 08:52:21 crc kubenswrapper[4810]: I0930 08:52:21.707527 4810 scope.go:117] "RemoveContainer" containerID="a274e25cded7421e038e2251ec27af7dbe8172d4c7e4d2c06168aa669836f3d0" Sep 30 08:52:21 crc kubenswrapper[4810]: I0930 08:52:21.755224 4810 scope.go:117] "RemoveContainer" containerID="612a628dc5694fefa9476ba97ed57198bc80e617d9dbe0dcfb09e14d0b996d8d" Sep 30 08:52:21 crc kubenswrapper[4810]: I0930 08:52:21.775474 4810 scope.go:117] "RemoveContainer" containerID="66e23df6841865f3e52782074053a40ffa2afa2c03e76c69473fa0655c6d52f3" Sep 30 08:52:31 crc kubenswrapper[4810]: I0930 08:52:31.318157 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:31 crc kubenswrapper[4810]: I0930 08:52:31.323651 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:32 crc kubenswrapper[4810]: I0930 08:52:32.250040 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.208415 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.210428 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.214974 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.215549 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vms5j" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.215601 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.217092 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.226651 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.298781 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.298876 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-config-data\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.299087 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.401498 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.402099 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl7hm\" (UniqueName: \"kubernetes.io/projected/78242ee7-a6aa-4087-832e-4834a7f4751f-kube-api-access-bl7hm\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.402212 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.402367 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-config-data\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.402601 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.403027 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.403253 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.403520 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.403767 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.403741 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.403812 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-config-data\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.414564 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.506713 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.507392 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.507752 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.507910 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl7hm\" (UniqueName: \"kubernetes.io/projected/78242ee7-a6aa-4087-832e-4834a7f4751f-kube-api-access-bl7hm\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.508051 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.508166 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.508249 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.508718 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.508995 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.515606 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.515704 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.539028 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl7hm\" (UniqueName: \"kubernetes.io/projected/78242ee7-a6aa-4087-832e-4834a7f4751f-kube-api-access-bl7hm\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.565131 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " pod="openstack/tempest-tests-tempest" Sep 30 08:52:48 crc kubenswrapper[4810]: I0930 08:52:48.847745 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 08:52:49 crc kubenswrapper[4810]: I0930 08:52:49.356838 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 08:52:49 crc kubenswrapper[4810]: I0930 08:52:49.358324 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Sep 30 08:52:49 crc kubenswrapper[4810]: I0930 08:52:49.426431 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"78242ee7-a6aa-4087-832e-4834a7f4751f","Type":"ContainerStarted","Data":"e30963a5616e7727f82bd8527dbc09b62c422e4faae0b2eaf8d32e5655813bf8"} Sep 30 08:52:58 crc kubenswrapper[4810]: I0930 08:52:58.359146 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Sep 30 08:52:59 crc kubenswrapper[4810]: I0930 08:52:59.563701 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"78242ee7-a6aa-4087-832e-4834a7f4751f","Type":"ContainerStarted","Data":"70d98fa1bc91e767b2be403c958203e9f54e59dd68eb8aefaa7cb90743a0f8ae"} Sep 30 08:52:59 crc kubenswrapper[4810]: I0930 08:52:59.596340 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.5985393549999998 podStartE2EDuration="12.596317232s" podCreationTimestamp="2025-09-30 08:52:47 +0000 UTC" firstStartedPulling="2025-09-30 08:52:49.356468302 +0000 UTC m=+2992.808667579" lastFinishedPulling="2025-09-30 08:52:58.354246179 +0000 UTC m=+3001.806445456" observedRunningTime="2025-09-30 08:52:59.591581805 +0000 UTC m=+3003.043781112" watchObservedRunningTime="2025-09-30 08:52:59.596317232 +0000 UTC m=+3003.048516499" Sep 30 08:54:15 crc kubenswrapper[4810]: I0930 08:54:15.911614 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:54:15 crc kubenswrapper[4810]: I0930 08:54:15.912459 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:54:45 crc kubenswrapper[4810]: I0930 08:54:45.912147 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:54:45 crc kubenswrapper[4810]: I0930 08:54:45.912854 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:55:15 crc kubenswrapper[4810]: I0930 08:55:15.911648 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 08:55:15 crc kubenswrapper[4810]: I0930 08:55:15.912179 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 08:55:15 crc kubenswrapper[4810]: I0930 08:55:15.912289 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 08:55:15 crc kubenswrapper[4810]: I0930 08:55:15.913174 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 08:55:15 crc kubenswrapper[4810]: I0930 08:55:15.913376 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" gracePeriod=600 Sep 30 08:55:16 crc kubenswrapper[4810]: E0930 08:55:16.040013 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:55:16 crc kubenswrapper[4810]: I0930 08:55:16.165034 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" exitCode=0 Sep 30 08:55:16 crc kubenswrapper[4810]: I0930 08:55:16.165077 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5"} Sep 30 08:55:16 crc kubenswrapper[4810]: I0930 08:55:16.165163 4810 scope.go:117] "RemoveContainer" containerID="2eb7a078098a0ac0ba80efe6dd4de079b7067414364ed78b82b35d53152b34ea" Sep 30 08:55:16 crc kubenswrapper[4810]: I0930 08:55:16.167510 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:55:16 crc kubenswrapper[4810]: E0930 08:55:16.167857 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:55:31 crc kubenswrapper[4810]: I0930 08:55:31.320455 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:55:31 crc kubenswrapper[4810]: E0930 08:55:31.321376 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:55:44 crc kubenswrapper[4810]: I0930 08:55:44.306334 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:55:44 crc kubenswrapper[4810]: E0930 08:55:44.307478 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:55:59 crc kubenswrapper[4810]: I0930 08:55:59.305817 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:55:59 crc kubenswrapper[4810]: E0930 08:55:59.306500 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.661635 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-54hb9"] Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.664829 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.671924 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-54hb9"] Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.805757 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-catalog-content\") pod \"redhat-marketplace-54hb9\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.806081 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcrxj\" (UniqueName: \"kubernetes.io/projected/3dedc6f0-1500-4138-a617-c76b64e4de04-kube-api-access-qcrxj\") pod \"redhat-marketplace-54hb9\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.806224 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-utilities\") pod \"redhat-marketplace-54hb9\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.907913 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcrxj\" (UniqueName: \"kubernetes.io/projected/3dedc6f0-1500-4138-a617-c76b64e4de04-kube-api-access-qcrxj\") pod \"redhat-marketplace-54hb9\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.907995 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-utilities\") pod \"redhat-marketplace-54hb9\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.908070 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-catalog-content\") pod \"redhat-marketplace-54hb9\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.908506 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-catalog-content\") pod \"redhat-marketplace-54hb9\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.908574 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-utilities\") pod \"redhat-marketplace-54hb9\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.936248 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcrxj\" (UniqueName: \"kubernetes.io/projected/3dedc6f0-1500-4138-a617-c76b64e4de04-kube-api-access-qcrxj\") pod \"redhat-marketplace-54hb9\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:13 crc kubenswrapper[4810]: I0930 08:56:13.987295 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:14 crc kubenswrapper[4810]: I0930 08:56:14.306827 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:56:14 crc kubenswrapper[4810]: E0930 08:56:14.307351 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:56:14 crc kubenswrapper[4810]: I0930 08:56:14.437916 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-54hb9"] Sep 30 08:56:14 crc kubenswrapper[4810]: I0930 08:56:14.814174 4810 generic.go:334] "Generic (PLEG): container finished" podID="3dedc6f0-1500-4138-a617-c76b64e4de04" containerID="2f48108ef85175933d1962ac966fe2e5d72c872f710a5a2ce9dc604d29991497" exitCode=0 Sep 30 08:56:14 crc kubenswrapper[4810]: I0930 08:56:14.815445 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54hb9" event={"ID":"3dedc6f0-1500-4138-a617-c76b64e4de04","Type":"ContainerDied","Data":"2f48108ef85175933d1962ac966fe2e5d72c872f710a5a2ce9dc604d29991497"} Sep 30 08:56:14 crc kubenswrapper[4810]: I0930 08:56:14.815616 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54hb9" event={"ID":"3dedc6f0-1500-4138-a617-c76b64e4de04","Type":"ContainerStarted","Data":"8fca35e2c13fdeadf729e47f7229231a80fd4f96ff11497cbddd821185349be8"} Sep 30 08:56:15 crc kubenswrapper[4810]: I0930 08:56:15.826811 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54hb9" event={"ID":"3dedc6f0-1500-4138-a617-c76b64e4de04","Type":"ContainerStarted","Data":"ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3"} Sep 30 08:56:16 crc kubenswrapper[4810]: I0930 08:56:16.840432 4810 generic.go:334] "Generic (PLEG): container finished" podID="3dedc6f0-1500-4138-a617-c76b64e4de04" containerID="ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3" exitCode=0 Sep 30 08:56:16 crc kubenswrapper[4810]: I0930 08:56:16.840554 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54hb9" event={"ID":"3dedc6f0-1500-4138-a617-c76b64e4de04","Type":"ContainerDied","Data":"ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3"} Sep 30 08:56:17 crc kubenswrapper[4810]: I0930 08:56:17.854400 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54hb9" event={"ID":"3dedc6f0-1500-4138-a617-c76b64e4de04","Type":"ContainerStarted","Data":"51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2"} Sep 30 08:56:17 crc kubenswrapper[4810]: I0930 08:56:17.887753 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-54hb9" podStartSLOduration=2.189053996 podStartE2EDuration="4.88773717s" podCreationTimestamp="2025-09-30 08:56:13 +0000 UTC" firstStartedPulling="2025-09-30 08:56:14.825718549 +0000 UTC m=+3198.277917806" lastFinishedPulling="2025-09-30 08:56:17.524401713 +0000 UTC m=+3200.976600980" observedRunningTime="2025-09-30 08:56:17.887170534 +0000 UTC m=+3201.339369821" watchObservedRunningTime="2025-09-30 08:56:17.88773717 +0000 UTC m=+3201.339936437" Sep 30 08:56:23 crc kubenswrapper[4810]: I0930 08:56:23.987609 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:23 crc kubenswrapper[4810]: I0930 08:56:23.988214 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:24 crc kubenswrapper[4810]: I0930 08:56:24.062621 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:25 crc kubenswrapper[4810]: I0930 08:56:25.021649 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:25 crc kubenswrapper[4810]: I0930 08:56:25.102390 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-54hb9"] Sep 30 08:56:26 crc kubenswrapper[4810]: I0930 08:56:26.963437 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-54hb9" podUID="3dedc6f0-1500-4138-a617-c76b64e4de04" containerName="registry-server" containerID="cri-o://51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2" gracePeriod=2 Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.308850 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:56:27 crc kubenswrapper[4810]: E0930 08:56:27.309892 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.505708 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.602901 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-catalog-content\") pod \"3dedc6f0-1500-4138-a617-c76b64e4de04\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.602968 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-utilities\") pod \"3dedc6f0-1500-4138-a617-c76b64e4de04\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.602993 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcrxj\" (UniqueName: \"kubernetes.io/projected/3dedc6f0-1500-4138-a617-c76b64e4de04-kube-api-access-qcrxj\") pod \"3dedc6f0-1500-4138-a617-c76b64e4de04\" (UID: \"3dedc6f0-1500-4138-a617-c76b64e4de04\") " Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.604108 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-utilities" (OuterVolumeSpecName: "utilities") pod "3dedc6f0-1500-4138-a617-c76b64e4de04" (UID: "3dedc6f0-1500-4138-a617-c76b64e4de04"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.608560 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dedc6f0-1500-4138-a617-c76b64e4de04-kube-api-access-qcrxj" (OuterVolumeSpecName: "kube-api-access-qcrxj") pod "3dedc6f0-1500-4138-a617-c76b64e4de04" (UID: "3dedc6f0-1500-4138-a617-c76b64e4de04"). InnerVolumeSpecName "kube-api-access-qcrxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.615083 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3dedc6f0-1500-4138-a617-c76b64e4de04" (UID: "3dedc6f0-1500-4138-a617-c76b64e4de04"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.705845 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.705886 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dedc6f0-1500-4138-a617-c76b64e4de04-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.705901 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcrxj\" (UniqueName: \"kubernetes.io/projected/3dedc6f0-1500-4138-a617-c76b64e4de04-kube-api-access-qcrxj\") on node \"crc\" DevicePath \"\"" Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.977183 4810 generic.go:334] "Generic (PLEG): container finished" podID="3dedc6f0-1500-4138-a617-c76b64e4de04" containerID="51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2" exitCode=0 Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.977241 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54hb9" Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.977248 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54hb9" event={"ID":"3dedc6f0-1500-4138-a617-c76b64e4de04","Type":"ContainerDied","Data":"51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2"} Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.978500 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54hb9" event={"ID":"3dedc6f0-1500-4138-a617-c76b64e4de04","Type":"ContainerDied","Data":"8fca35e2c13fdeadf729e47f7229231a80fd4f96ff11497cbddd821185349be8"} Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.978551 4810 scope.go:117] "RemoveContainer" containerID="51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2" Sep 30 08:56:27 crc kubenswrapper[4810]: I0930 08:56:27.998593 4810 scope.go:117] "RemoveContainer" containerID="ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3" Sep 30 08:56:28 crc kubenswrapper[4810]: I0930 08:56:28.028640 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-54hb9"] Sep 30 08:56:28 crc kubenswrapper[4810]: I0930 08:56:28.043373 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-54hb9"] Sep 30 08:56:28 crc kubenswrapper[4810]: I0930 08:56:28.055599 4810 scope.go:117] "RemoveContainer" containerID="2f48108ef85175933d1962ac966fe2e5d72c872f710a5a2ce9dc604d29991497" Sep 30 08:56:28 crc kubenswrapper[4810]: I0930 08:56:28.099427 4810 scope.go:117] "RemoveContainer" containerID="51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2" Sep 30 08:56:28 crc kubenswrapper[4810]: E0930 08:56:28.100057 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2\": container with ID starting with 51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2 not found: ID does not exist" containerID="51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2" Sep 30 08:56:28 crc kubenswrapper[4810]: I0930 08:56:28.100121 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2"} err="failed to get container status \"51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2\": rpc error: code = NotFound desc = could not find container \"51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2\": container with ID starting with 51846527d7f7aa692da4ac3fda742751ae7416f1eed3d6d0858f12682b8bece2 not found: ID does not exist" Sep 30 08:56:28 crc kubenswrapper[4810]: I0930 08:56:28.100150 4810 scope.go:117] "RemoveContainer" containerID="ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3" Sep 30 08:56:28 crc kubenswrapper[4810]: E0930 08:56:28.100913 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3\": container with ID starting with ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3 not found: ID does not exist" containerID="ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3" Sep 30 08:56:28 crc kubenswrapper[4810]: I0930 08:56:28.101018 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3"} err="failed to get container status \"ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3\": rpc error: code = NotFound desc = could not find container \"ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3\": container with ID starting with ac9ccc80c3ff8f5a5e68166e2529a43fbd7a55c8cea29243740c726ba77c7cc3 not found: ID does not exist" Sep 30 08:56:28 crc kubenswrapper[4810]: I0930 08:56:28.101116 4810 scope.go:117] "RemoveContainer" containerID="2f48108ef85175933d1962ac966fe2e5d72c872f710a5a2ce9dc604d29991497" Sep 30 08:56:28 crc kubenswrapper[4810]: E0930 08:56:28.101490 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f48108ef85175933d1962ac966fe2e5d72c872f710a5a2ce9dc604d29991497\": container with ID starting with 2f48108ef85175933d1962ac966fe2e5d72c872f710a5a2ce9dc604d29991497 not found: ID does not exist" containerID="2f48108ef85175933d1962ac966fe2e5d72c872f710a5a2ce9dc604d29991497" Sep 30 08:56:28 crc kubenswrapper[4810]: I0930 08:56:28.101533 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f48108ef85175933d1962ac966fe2e5d72c872f710a5a2ce9dc604d29991497"} err="failed to get container status \"2f48108ef85175933d1962ac966fe2e5d72c872f710a5a2ce9dc604d29991497\": rpc error: code = NotFound desc = could not find container \"2f48108ef85175933d1962ac966fe2e5d72c872f710a5a2ce9dc604d29991497\": container with ID starting with 2f48108ef85175933d1962ac966fe2e5d72c872f710a5a2ce9dc604d29991497 not found: ID does not exist" Sep 30 08:56:29 crc kubenswrapper[4810]: I0930 08:56:29.321050 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dedc6f0-1500-4138-a617-c76b64e4de04" path="/var/lib/kubelet/pods/3dedc6f0-1500-4138-a617-c76b64e4de04/volumes" Sep 30 08:56:42 crc kubenswrapper[4810]: I0930 08:56:42.306372 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:56:42 crc kubenswrapper[4810]: E0930 08:56:42.309674 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:56:54 crc kubenswrapper[4810]: I0930 08:56:54.307518 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:56:54 crc kubenswrapper[4810]: E0930 08:56:54.308465 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:57:08 crc kubenswrapper[4810]: I0930 08:57:08.306969 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:57:08 crc kubenswrapper[4810]: E0930 08:57:08.308083 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:57:22 crc kubenswrapper[4810]: I0930 08:57:22.306750 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:57:22 crc kubenswrapper[4810]: E0930 08:57:22.307891 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:57:35 crc kubenswrapper[4810]: I0930 08:57:35.307223 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:57:35 crc kubenswrapper[4810]: E0930 08:57:35.308214 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.650355 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nq7k8"] Sep 30 08:57:39 crc kubenswrapper[4810]: E0930 08:57:39.651227 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dedc6f0-1500-4138-a617-c76b64e4de04" containerName="extract-content" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.651238 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dedc6f0-1500-4138-a617-c76b64e4de04" containerName="extract-content" Sep 30 08:57:39 crc kubenswrapper[4810]: E0930 08:57:39.651258 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dedc6f0-1500-4138-a617-c76b64e4de04" containerName="registry-server" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.651301 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dedc6f0-1500-4138-a617-c76b64e4de04" containerName="registry-server" Sep 30 08:57:39 crc kubenswrapper[4810]: E0930 08:57:39.651323 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dedc6f0-1500-4138-a617-c76b64e4de04" containerName="extract-utilities" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.651330 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dedc6f0-1500-4138-a617-c76b64e4de04" containerName="extract-utilities" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.651528 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dedc6f0-1500-4138-a617-c76b64e4de04" containerName="registry-server" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.653086 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.671053 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nq7k8"] Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.735679 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tgdk\" (UniqueName: \"kubernetes.io/projected/91fdc5ad-80ad-43fe-989c-178670a07dd5-kube-api-access-4tgdk\") pod \"certified-operators-nq7k8\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.736024 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-utilities\") pod \"certified-operators-nq7k8\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.736117 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-catalog-content\") pod \"certified-operators-nq7k8\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.838796 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tgdk\" (UniqueName: \"kubernetes.io/projected/91fdc5ad-80ad-43fe-989c-178670a07dd5-kube-api-access-4tgdk\") pod \"certified-operators-nq7k8\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.838888 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-utilities\") pod \"certified-operators-nq7k8\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.838917 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-catalog-content\") pod \"certified-operators-nq7k8\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.839532 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-catalog-content\") pod \"certified-operators-nq7k8\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.839848 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-utilities\") pod \"certified-operators-nq7k8\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.871423 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tgdk\" (UniqueName: \"kubernetes.io/projected/91fdc5ad-80ad-43fe-989c-178670a07dd5-kube-api-access-4tgdk\") pod \"certified-operators-nq7k8\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:39 crc kubenswrapper[4810]: I0930 08:57:39.972252 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:40 crc kubenswrapper[4810]: I0930 08:57:40.498153 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nq7k8"] Sep 30 08:57:40 crc kubenswrapper[4810]: I0930 08:57:40.878115 4810 generic.go:334] "Generic (PLEG): container finished" podID="91fdc5ad-80ad-43fe-989c-178670a07dd5" containerID="623e112949ea5f19c8e5cac648903fa21bf49bd05b326ff56e79ee978f51d7c5" exitCode=0 Sep 30 08:57:40 crc kubenswrapper[4810]: I0930 08:57:40.878167 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nq7k8" event={"ID":"91fdc5ad-80ad-43fe-989c-178670a07dd5","Type":"ContainerDied","Data":"623e112949ea5f19c8e5cac648903fa21bf49bd05b326ff56e79ee978f51d7c5"} Sep 30 08:57:40 crc kubenswrapper[4810]: I0930 08:57:40.878199 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nq7k8" event={"ID":"91fdc5ad-80ad-43fe-989c-178670a07dd5","Type":"ContainerStarted","Data":"a051c243b2bf93fb0c2d59dbd028b346394349486d280c296588dc81aa6a22c8"} Sep 30 08:57:41 crc kubenswrapper[4810]: I0930 08:57:41.891471 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nq7k8" event={"ID":"91fdc5ad-80ad-43fe-989c-178670a07dd5","Type":"ContainerStarted","Data":"32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11"} Sep 30 08:57:42 crc kubenswrapper[4810]: I0930 08:57:42.905246 4810 generic.go:334] "Generic (PLEG): container finished" podID="91fdc5ad-80ad-43fe-989c-178670a07dd5" containerID="32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11" exitCode=0 Sep 30 08:57:42 crc kubenswrapper[4810]: I0930 08:57:42.905404 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nq7k8" event={"ID":"91fdc5ad-80ad-43fe-989c-178670a07dd5","Type":"ContainerDied","Data":"32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11"} Sep 30 08:57:43 crc kubenswrapper[4810]: I0930 08:57:43.921293 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nq7k8" event={"ID":"91fdc5ad-80ad-43fe-989c-178670a07dd5","Type":"ContainerStarted","Data":"cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7"} Sep 30 08:57:43 crc kubenswrapper[4810]: I0930 08:57:43.957331 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nq7k8" podStartSLOduration=2.418897979 podStartE2EDuration="4.957312828s" podCreationTimestamp="2025-09-30 08:57:39 +0000 UTC" firstStartedPulling="2025-09-30 08:57:40.880529288 +0000 UTC m=+3284.332728555" lastFinishedPulling="2025-09-30 08:57:43.418944137 +0000 UTC m=+3286.871143404" observedRunningTime="2025-09-30 08:57:43.953565934 +0000 UTC m=+3287.405765221" watchObservedRunningTime="2025-09-30 08:57:43.957312828 +0000 UTC m=+3287.409512095" Sep 30 08:57:46 crc kubenswrapper[4810]: I0930 08:57:46.310459 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:57:46 crc kubenswrapper[4810]: E0930 08:57:46.311563 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:57:49 crc kubenswrapper[4810]: I0930 08:57:49.972391 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:49 crc kubenswrapper[4810]: I0930 08:57:49.972989 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:50 crc kubenswrapper[4810]: I0930 08:57:50.027596 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:50 crc kubenswrapper[4810]: I0930 08:57:50.081214 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:50 crc kubenswrapper[4810]: I0930 08:57:50.265414 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nq7k8"] Sep 30 08:57:52 crc kubenswrapper[4810]: I0930 08:57:52.023181 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nq7k8" podUID="91fdc5ad-80ad-43fe-989c-178670a07dd5" containerName="registry-server" containerID="cri-o://cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7" gracePeriod=2 Sep 30 08:57:52 crc kubenswrapper[4810]: I0930 08:57:52.564847 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:52 crc kubenswrapper[4810]: I0930 08:57:52.719461 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-utilities\") pod \"91fdc5ad-80ad-43fe-989c-178670a07dd5\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " Sep 30 08:57:52 crc kubenswrapper[4810]: I0930 08:57:52.719632 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tgdk\" (UniqueName: \"kubernetes.io/projected/91fdc5ad-80ad-43fe-989c-178670a07dd5-kube-api-access-4tgdk\") pod \"91fdc5ad-80ad-43fe-989c-178670a07dd5\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " Sep 30 08:57:52 crc kubenswrapper[4810]: I0930 08:57:52.719726 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-catalog-content\") pod \"91fdc5ad-80ad-43fe-989c-178670a07dd5\" (UID: \"91fdc5ad-80ad-43fe-989c-178670a07dd5\") " Sep 30 08:57:52 crc kubenswrapper[4810]: I0930 08:57:52.720559 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-utilities" (OuterVolumeSpecName: "utilities") pod "91fdc5ad-80ad-43fe-989c-178670a07dd5" (UID: "91fdc5ad-80ad-43fe-989c-178670a07dd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:57:52 crc kubenswrapper[4810]: I0930 08:57:52.732587 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91fdc5ad-80ad-43fe-989c-178670a07dd5-kube-api-access-4tgdk" (OuterVolumeSpecName: "kube-api-access-4tgdk") pod "91fdc5ad-80ad-43fe-989c-178670a07dd5" (UID: "91fdc5ad-80ad-43fe-989c-178670a07dd5"). InnerVolumeSpecName "kube-api-access-4tgdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:57:52 crc kubenswrapper[4810]: I0930 08:57:52.770738 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91fdc5ad-80ad-43fe-989c-178670a07dd5" (UID: "91fdc5ad-80ad-43fe-989c-178670a07dd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:57:52 crc kubenswrapper[4810]: I0930 08:57:52.822161 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:57:52 crc kubenswrapper[4810]: I0930 08:57:52.822208 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tgdk\" (UniqueName: \"kubernetes.io/projected/91fdc5ad-80ad-43fe-989c-178670a07dd5-kube-api-access-4tgdk\") on node \"crc\" DevicePath \"\"" Sep 30 08:57:52 crc kubenswrapper[4810]: I0930 08:57:52.822224 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91fdc5ad-80ad-43fe-989c-178670a07dd5-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.036234 4810 generic.go:334] "Generic (PLEG): container finished" podID="91fdc5ad-80ad-43fe-989c-178670a07dd5" containerID="cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7" exitCode=0 Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.036325 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nq7k8" event={"ID":"91fdc5ad-80ad-43fe-989c-178670a07dd5","Type":"ContainerDied","Data":"cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7"} Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.036346 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nq7k8" Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.036368 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nq7k8" event={"ID":"91fdc5ad-80ad-43fe-989c-178670a07dd5","Type":"ContainerDied","Data":"a051c243b2bf93fb0c2d59dbd028b346394349486d280c296588dc81aa6a22c8"} Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.036399 4810 scope.go:117] "RemoveContainer" containerID="cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7" Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.075522 4810 scope.go:117] "RemoveContainer" containerID="32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11" Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.079181 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nq7k8"] Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.087897 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nq7k8"] Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.111427 4810 scope.go:117] "RemoveContainer" containerID="623e112949ea5f19c8e5cac648903fa21bf49bd05b326ff56e79ee978f51d7c5" Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.191295 4810 scope.go:117] "RemoveContainer" containerID="cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7" Sep 30 08:57:53 crc kubenswrapper[4810]: E0930 08:57:53.192142 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7\": container with ID starting with cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7 not found: ID does not exist" containerID="cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7" Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.192174 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7"} err="failed to get container status \"cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7\": rpc error: code = NotFound desc = could not find container \"cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7\": container with ID starting with cef8d07ac053419b029ee903e643deaeb4977719b0e443d302fb1918682532c7 not found: ID does not exist" Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.192196 4810 scope.go:117] "RemoveContainer" containerID="32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11" Sep 30 08:57:53 crc kubenswrapper[4810]: E0930 08:57:53.192824 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11\": container with ID starting with 32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11 not found: ID does not exist" containerID="32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11" Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.192869 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11"} err="failed to get container status \"32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11\": rpc error: code = NotFound desc = could not find container \"32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11\": container with ID starting with 32cb8162f8803be87dcd164afc129014b9637b584a19b80b243743a0f338ac11 not found: ID does not exist" Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.192882 4810 scope.go:117] "RemoveContainer" containerID="623e112949ea5f19c8e5cac648903fa21bf49bd05b326ff56e79ee978f51d7c5" Sep 30 08:57:53 crc kubenswrapper[4810]: E0930 08:57:53.193153 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"623e112949ea5f19c8e5cac648903fa21bf49bd05b326ff56e79ee978f51d7c5\": container with ID starting with 623e112949ea5f19c8e5cac648903fa21bf49bd05b326ff56e79ee978f51d7c5 not found: ID does not exist" containerID="623e112949ea5f19c8e5cac648903fa21bf49bd05b326ff56e79ee978f51d7c5" Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.193183 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"623e112949ea5f19c8e5cac648903fa21bf49bd05b326ff56e79ee978f51d7c5"} err="failed to get container status \"623e112949ea5f19c8e5cac648903fa21bf49bd05b326ff56e79ee978f51d7c5\": rpc error: code = NotFound desc = could not find container \"623e112949ea5f19c8e5cac648903fa21bf49bd05b326ff56e79ee978f51d7c5\": container with ID starting with 623e112949ea5f19c8e5cac648903fa21bf49bd05b326ff56e79ee978f51d7c5 not found: ID does not exist" Sep 30 08:57:53 crc kubenswrapper[4810]: I0930 08:57:53.318998 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91fdc5ad-80ad-43fe-989c-178670a07dd5" path="/var/lib/kubelet/pods/91fdc5ad-80ad-43fe-989c-178670a07dd5/volumes" Sep 30 08:58:01 crc kubenswrapper[4810]: I0930 08:58:01.307489 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:58:01 crc kubenswrapper[4810]: E0930 08:58:01.308557 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:58:12 crc kubenswrapper[4810]: I0930 08:58:12.307000 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:58:12 crc kubenswrapper[4810]: E0930 08:58:12.307646 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:58:25 crc kubenswrapper[4810]: I0930 08:58:25.306129 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:58:25 crc kubenswrapper[4810]: E0930 08:58:25.306842 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:58:36 crc kubenswrapper[4810]: I0930 08:58:36.306852 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:58:36 crc kubenswrapper[4810]: E0930 08:58:36.307619 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:58:50 crc kubenswrapper[4810]: I0930 08:58:50.306821 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:58:50 crc kubenswrapper[4810]: E0930 08:58:50.307536 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:58:59 crc kubenswrapper[4810]: I0930 08:58:59.873405 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cmkf7"] Sep 30 08:58:59 crc kubenswrapper[4810]: E0930 08:58:59.874425 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91fdc5ad-80ad-43fe-989c-178670a07dd5" containerName="registry-server" Sep 30 08:58:59 crc kubenswrapper[4810]: I0930 08:58:59.874438 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="91fdc5ad-80ad-43fe-989c-178670a07dd5" containerName="registry-server" Sep 30 08:58:59 crc kubenswrapper[4810]: E0930 08:58:59.874460 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91fdc5ad-80ad-43fe-989c-178670a07dd5" containerName="extract-content" Sep 30 08:58:59 crc kubenswrapper[4810]: I0930 08:58:59.874466 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="91fdc5ad-80ad-43fe-989c-178670a07dd5" containerName="extract-content" Sep 30 08:58:59 crc kubenswrapper[4810]: E0930 08:58:59.874476 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91fdc5ad-80ad-43fe-989c-178670a07dd5" containerName="extract-utilities" Sep 30 08:58:59 crc kubenswrapper[4810]: I0930 08:58:59.874483 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="91fdc5ad-80ad-43fe-989c-178670a07dd5" containerName="extract-utilities" Sep 30 08:58:59 crc kubenswrapper[4810]: I0930 08:58:59.874664 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="91fdc5ad-80ad-43fe-989c-178670a07dd5" containerName="registry-server" Sep 30 08:58:59 crc kubenswrapper[4810]: I0930 08:58:59.876079 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:58:59 crc kubenswrapper[4810]: I0930 08:58:59.885503 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cmkf7"] Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.026539 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-catalog-content\") pod \"redhat-operators-cmkf7\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.026891 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-utilities\") pod \"redhat-operators-cmkf7\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.027145 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tnhv\" (UniqueName: \"kubernetes.io/projected/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-kube-api-access-9tnhv\") pod \"redhat-operators-cmkf7\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.128903 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-catalog-content\") pod \"redhat-operators-cmkf7\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.128990 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-utilities\") pod \"redhat-operators-cmkf7\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.129053 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tnhv\" (UniqueName: \"kubernetes.io/projected/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-kube-api-access-9tnhv\") pod \"redhat-operators-cmkf7\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.129553 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-catalog-content\") pod \"redhat-operators-cmkf7\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.129739 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-utilities\") pod \"redhat-operators-cmkf7\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.147669 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tnhv\" (UniqueName: \"kubernetes.io/projected/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-kube-api-access-9tnhv\") pod \"redhat-operators-cmkf7\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.208952 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.706973 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cmkf7"] Sep 30 08:59:00 crc kubenswrapper[4810]: I0930 08:59:00.750015 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmkf7" event={"ID":"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9","Type":"ContainerStarted","Data":"45af86b2258e1160d979ce61adb311ed9aa873afd83bbf636ed03cd18e4187a3"} Sep 30 08:59:01 crc kubenswrapper[4810]: I0930 08:59:01.306725 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:59:01 crc kubenswrapper[4810]: E0930 08:59:01.307038 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:59:01 crc kubenswrapper[4810]: I0930 08:59:01.763545 4810 generic.go:334] "Generic (PLEG): container finished" podID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerID="e04ce651e51fada46fe4941abf56f88ac495768b9bfd9d663ca8b4f1d4e87d14" exitCode=0 Sep 30 08:59:01 crc kubenswrapper[4810]: I0930 08:59:01.763599 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmkf7" event={"ID":"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9","Type":"ContainerDied","Data":"e04ce651e51fada46fe4941abf56f88ac495768b9bfd9d663ca8b4f1d4e87d14"} Sep 30 08:59:01 crc kubenswrapper[4810]: I0930 08:59:01.766390 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 08:59:02 crc kubenswrapper[4810]: I0930 08:59:02.778085 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmkf7" event={"ID":"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9","Type":"ContainerStarted","Data":"7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec"} Sep 30 08:59:05 crc kubenswrapper[4810]: I0930 08:59:05.815188 4810 generic.go:334] "Generic (PLEG): container finished" podID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerID="7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec" exitCode=0 Sep 30 08:59:05 crc kubenswrapper[4810]: I0930 08:59:05.815257 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmkf7" event={"ID":"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9","Type":"ContainerDied","Data":"7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec"} Sep 30 08:59:06 crc kubenswrapper[4810]: I0930 08:59:06.830078 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmkf7" event={"ID":"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9","Type":"ContainerStarted","Data":"16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b"} Sep 30 08:59:06 crc kubenswrapper[4810]: I0930 08:59:06.857490 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cmkf7" podStartSLOduration=3.101586939 podStartE2EDuration="7.857466393s" podCreationTimestamp="2025-09-30 08:58:59 +0000 UTC" firstStartedPulling="2025-09-30 08:59:01.765885407 +0000 UTC m=+3365.218084704" lastFinishedPulling="2025-09-30 08:59:06.521764841 +0000 UTC m=+3369.973964158" observedRunningTime="2025-09-30 08:59:06.854347912 +0000 UTC m=+3370.306547189" watchObservedRunningTime="2025-09-30 08:59:06.857466393 +0000 UTC m=+3370.309665670" Sep 30 08:59:10 crc kubenswrapper[4810]: I0930 08:59:10.209439 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:10 crc kubenswrapper[4810]: I0930 08:59:10.209725 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:11 crc kubenswrapper[4810]: I0930 08:59:11.275358 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cmkf7" podUID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerName="registry-server" probeResult="failure" output=< Sep 30 08:59:11 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Sep 30 08:59:11 crc kubenswrapper[4810]: > Sep 30 08:59:16 crc kubenswrapper[4810]: I0930 08:59:16.307901 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:59:16 crc kubenswrapper[4810]: E0930 08:59:16.309022 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:59:21 crc kubenswrapper[4810]: I0930 08:59:21.276760 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cmkf7" podUID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerName="registry-server" probeResult="failure" output=< Sep 30 08:59:21 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Sep 30 08:59:21 crc kubenswrapper[4810]: > Sep 30 08:59:30 crc kubenswrapper[4810]: I0930 08:59:30.288539 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:30 crc kubenswrapper[4810]: I0930 08:59:30.343629 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:31 crc kubenswrapper[4810]: I0930 08:59:31.063024 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cmkf7"] Sep 30 08:59:31 crc kubenswrapper[4810]: I0930 08:59:31.306729 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:59:31 crc kubenswrapper[4810]: E0930 08:59:31.307058 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:59:32 crc kubenswrapper[4810]: I0930 08:59:32.087847 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cmkf7" podUID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerName="registry-server" containerID="cri-o://16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b" gracePeriod=2 Sep 30 08:59:32 crc kubenswrapper[4810]: I0930 08:59:32.513492 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:32 crc kubenswrapper[4810]: I0930 08:59:32.553560 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-utilities\") pod \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " Sep 30 08:59:32 crc kubenswrapper[4810]: I0930 08:59:32.553642 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tnhv\" (UniqueName: \"kubernetes.io/projected/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-kube-api-access-9tnhv\") pod \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " Sep 30 08:59:32 crc kubenswrapper[4810]: I0930 08:59:32.553729 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-catalog-content\") pod \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\" (UID: \"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9\") " Sep 30 08:59:32 crc kubenswrapper[4810]: I0930 08:59:32.554392 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-utilities" (OuterVolumeSpecName: "utilities") pod "a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" (UID: "a427fcc2-41a3-46dc-af7d-1fd5effcb1b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:59:32 crc kubenswrapper[4810]: I0930 08:59:32.562936 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-kube-api-access-9tnhv" (OuterVolumeSpecName: "kube-api-access-9tnhv") pod "a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" (UID: "a427fcc2-41a3-46dc-af7d-1fd5effcb1b9"). InnerVolumeSpecName "kube-api-access-9tnhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 08:59:32 crc kubenswrapper[4810]: I0930 08:59:32.656624 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 08:59:32 crc kubenswrapper[4810]: I0930 08:59:32.656857 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tnhv\" (UniqueName: \"kubernetes.io/projected/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-kube-api-access-9tnhv\") on node \"crc\" DevicePath \"\"" Sep 30 08:59:32 crc kubenswrapper[4810]: I0930 08:59:32.663809 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" (UID: "a427fcc2-41a3-46dc-af7d-1fd5effcb1b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 08:59:32 crc kubenswrapper[4810]: I0930 08:59:32.787566 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.104879 4810 generic.go:334] "Generic (PLEG): container finished" podID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerID="16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b" exitCode=0 Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.104955 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmkf7" event={"ID":"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9","Type":"ContainerDied","Data":"16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b"} Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.105026 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmkf7" event={"ID":"a427fcc2-41a3-46dc-af7d-1fd5effcb1b9","Type":"ContainerDied","Data":"45af86b2258e1160d979ce61adb311ed9aa873afd83bbf636ed03cd18e4187a3"} Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.105055 4810 scope.go:117] "RemoveContainer" containerID="16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b" Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.105853 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cmkf7" Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.160060 4810 scope.go:117] "RemoveContainer" containerID="7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec" Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.169338 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cmkf7"] Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.182735 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cmkf7"] Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.185052 4810 scope.go:117] "RemoveContainer" containerID="e04ce651e51fada46fe4941abf56f88ac495768b9bfd9d663ca8b4f1d4e87d14" Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.247124 4810 scope.go:117] "RemoveContainer" containerID="16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b" Sep 30 08:59:33 crc kubenswrapper[4810]: E0930 08:59:33.247538 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b\": container with ID starting with 16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b not found: ID does not exist" containerID="16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b" Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.247572 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b"} err="failed to get container status \"16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b\": rpc error: code = NotFound desc = could not find container \"16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b\": container with ID starting with 16c236ed59bc5c827cc524f15f7f159246ed3ec6da9f4b064f9fb2f4b1382e0b not found: ID does not exist" Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.247594 4810 scope.go:117] "RemoveContainer" containerID="7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec" Sep 30 08:59:33 crc kubenswrapper[4810]: E0930 08:59:33.247983 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec\": container with ID starting with 7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec not found: ID does not exist" containerID="7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec" Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.248014 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec"} err="failed to get container status \"7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec\": rpc error: code = NotFound desc = could not find container \"7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec\": container with ID starting with 7ba3d2a46a6ae5fcd24c5da613a45e7dbb9b8299cf80e1a4fd303ba9a1de5cec not found: ID does not exist" Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.248031 4810 scope.go:117] "RemoveContainer" containerID="e04ce651e51fada46fe4941abf56f88ac495768b9bfd9d663ca8b4f1d4e87d14" Sep 30 08:59:33 crc kubenswrapper[4810]: E0930 08:59:33.248283 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e04ce651e51fada46fe4941abf56f88ac495768b9bfd9d663ca8b4f1d4e87d14\": container with ID starting with e04ce651e51fada46fe4941abf56f88ac495768b9bfd9d663ca8b4f1d4e87d14 not found: ID does not exist" containerID="e04ce651e51fada46fe4941abf56f88ac495768b9bfd9d663ca8b4f1d4e87d14" Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.248315 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e04ce651e51fada46fe4941abf56f88ac495768b9bfd9d663ca8b4f1d4e87d14"} err="failed to get container status \"e04ce651e51fada46fe4941abf56f88ac495768b9bfd9d663ca8b4f1d4e87d14\": rpc error: code = NotFound desc = could not find container \"e04ce651e51fada46fe4941abf56f88ac495768b9bfd9d663ca8b4f1d4e87d14\": container with ID starting with e04ce651e51fada46fe4941abf56f88ac495768b9bfd9d663ca8b4f1d4e87d14 not found: ID does not exist" Sep 30 08:59:33 crc kubenswrapper[4810]: I0930 08:59:33.316868 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" path="/var/lib/kubelet/pods/a427fcc2-41a3-46dc-af7d-1fd5effcb1b9/volumes" Sep 30 08:59:43 crc kubenswrapper[4810]: I0930 08:59:43.307330 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:59:43 crc kubenswrapper[4810]: E0930 08:59:43.308390 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 08:59:54 crc kubenswrapper[4810]: I0930 08:59:54.307459 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 08:59:54 crc kubenswrapper[4810]: E0930 08:59:54.308347 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.192005 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h"] Sep 30 09:00:00 crc kubenswrapper[4810]: E0930 09:00:00.192994 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerName="registry-server" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.193009 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerName="registry-server" Sep 30 09:00:00 crc kubenswrapper[4810]: E0930 09:00:00.193028 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerName="extract-utilities" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.193038 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerName="extract-utilities" Sep 30 09:00:00 crc kubenswrapper[4810]: E0930 09:00:00.193052 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerName="extract-content" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.193060 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerName="extract-content" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.193345 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a427fcc2-41a3-46dc-af7d-1fd5effcb1b9" containerName="registry-server" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.194143 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.196733 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.196879 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.202658 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h"] Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.306629 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-config-volume\") pod \"collect-profiles-29320380-fkg4h\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.306816 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftrl5\" (UniqueName: \"kubernetes.io/projected/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-kube-api-access-ftrl5\") pod \"collect-profiles-29320380-fkg4h\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.306906 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-secret-volume\") pod \"collect-profiles-29320380-fkg4h\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.409557 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-config-volume\") pod \"collect-profiles-29320380-fkg4h\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.409784 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftrl5\" (UniqueName: \"kubernetes.io/projected/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-kube-api-access-ftrl5\") pod \"collect-profiles-29320380-fkg4h\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.409885 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-secret-volume\") pod \"collect-profiles-29320380-fkg4h\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.411012 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-config-volume\") pod \"collect-profiles-29320380-fkg4h\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.423299 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-secret-volume\") pod \"collect-profiles-29320380-fkg4h\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.430098 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftrl5\" (UniqueName: \"kubernetes.io/projected/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-kube-api-access-ftrl5\") pod \"collect-profiles-29320380-fkg4h\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.522755 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:00 crc kubenswrapper[4810]: I0930 09:00:00.988464 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h"] Sep 30 09:00:01 crc kubenswrapper[4810]: I0930 09:00:01.522731 4810 generic.go:334] "Generic (PLEG): container finished" podID="6d7cf5d7-ce91-4c1b-b025-2edb9e25353c" containerID="c28a79315907100a129348ab98b29ba1ed0e3503b4e473970cc164e2fcfd6b7d" exitCode=0 Sep 30 09:00:01 crc kubenswrapper[4810]: I0930 09:00:01.522979 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" event={"ID":"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c","Type":"ContainerDied","Data":"c28a79315907100a129348ab98b29ba1ed0e3503b4e473970cc164e2fcfd6b7d"} Sep 30 09:00:01 crc kubenswrapper[4810]: I0930 09:00:01.523002 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" event={"ID":"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c","Type":"ContainerStarted","Data":"c4871cd06fc6850478ac19f9c7ab69df771cea539217a8f74e4e75ef082c4780"} Sep 30 09:00:02 crc kubenswrapper[4810]: I0930 09:00:02.894210 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:02 crc kubenswrapper[4810]: I0930 09:00:02.959312 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-secret-volume\") pod \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " Sep 30 09:00:02 crc kubenswrapper[4810]: I0930 09:00:02.959437 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftrl5\" (UniqueName: \"kubernetes.io/projected/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-kube-api-access-ftrl5\") pod \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " Sep 30 09:00:02 crc kubenswrapper[4810]: I0930 09:00:02.959558 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-config-volume\") pod \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\" (UID: \"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c\") " Sep 30 09:00:02 crc kubenswrapper[4810]: I0930 09:00:02.960494 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-config-volume" (OuterVolumeSpecName: "config-volume") pod "6d7cf5d7-ce91-4c1b-b025-2edb9e25353c" (UID: "6d7cf5d7-ce91-4c1b-b025-2edb9e25353c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 09:00:02 crc kubenswrapper[4810]: I0930 09:00:02.965488 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-kube-api-access-ftrl5" (OuterVolumeSpecName: "kube-api-access-ftrl5") pod "6d7cf5d7-ce91-4c1b-b025-2edb9e25353c" (UID: "6d7cf5d7-ce91-4c1b-b025-2edb9e25353c"). InnerVolumeSpecName "kube-api-access-ftrl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:00:02 crc kubenswrapper[4810]: I0930 09:00:02.966087 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6d7cf5d7-ce91-4c1b-b025-2edb9e25353c" (UID: "6d7cf5d7-ce91-4c1b-b025-2edb9e25353c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 09:00:03 crc kubenswrapper[4810]: I0930 09:00:03.061376 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 09:00:03 crc kubenswrapper[4810]: I0930 09:00:03.061409 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 09:00:03 crc kubenswrapper[4810]: I0930 09:00:03.061419 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftrl5\" (UniqueName: \"kubernetes.io/projected/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c-kube-api-access-ftrl5\") on node \"crc\" DevicePath \"\"" Sep 30 09:00:03 crc kubenswrapper[4810]: I0930 09:00:03.545106 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" event={"ID":"6d7cf5d7-ce91-4c1b-b025-2edb9e25353c","Type":"ContainerDied","Data":"c4871cd06fc6850478ac19f9c7ab69df771cea539217a8f74e4e75ef082c4780"} Sep 30 09:00:03 crc kubenswrapper[4810]: I0930 09:00:03.545153 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4871cd06fc6850478ac19f9c7ab69df771cea539217a8f74e4e75ef082c4780" Sep 30 09:00:03 crc kubenswrapper[4810]: I0930 09:00:03.545193 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h" Sep 30 09:00:04 crc kubenswrapper[4810]: I0930 09:00:04.009490 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8"] Sep 30 09:00:04 crc kubenswrapper[4810]: I0930 09:00:04.019615 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320335-9wnd8"] Sep 30 09:00:05 crc kubenswrapper[4810]: I0930 09:00:05.330662 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ce63664-583b-408a-a9d1-061552873891" path="/var/lib/kubelet/pods/9ce63664-583b-408a-a9d1-061552873891/volumes" Sep 30 09:00:07 crc kubenswrapper[4810]: I0930 09:00:07.314838 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 09:00:07 crc kubenswrapper[4810]: E0930 09:00:07.315410 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:00:18 crc kubenswrapper[4810]: I0930 09:00:18.307317 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 09:00:18 crc kubenswrapper[4810]: I0930 09:00:18.709088 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"b04c10958aa5c72edbc42e5ad3647261d4e99fe293d352aeae3dd3f087658bc1"} Sep 30 09:00:22 crc kubenswrapper[4810]: I0930 09:00:22.027909 4810 scope.go:117] "RemoveContainer" containerID="4a8ae09deb8c59bb0dec3236ae6e7affa65036022541214f402cfa5168f3ddf1" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.663889 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5kk9g"] Sep 30 09:00:47 crc kubenswrapper[4810]: E0930 09:00:47.665133 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d7cf5d7-ce91-4c1b-b025-2edb9e25353c" containerName="collect-profiles" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.665155 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d7cf5d7-ce91-4c1b-b025-2edb9e25353c" containerName="collect-profiles" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.665560 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d7cf5d7-ce91-4c1b-b025-2edb9e25353c" containerName="collect-profiles" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.668387 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.733530 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5kk9g"] Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.814317 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-utilities\") pod \"community-operators-5kk9g\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.814430 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-catalog-content\") pod \"community-operators-5kk9g\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.814584 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br8th\" (UniqueName: \"kubernetes.io/projected/173e3771-e034-4856-ba30-9eecf64a83c6-kube-api-access-br8th\") pod \"community-operators-5kk9g\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.916052 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-catalog-content\") pod \"community-operators-5kk9g\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.916194 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br8th\" (UniqueName: \"kubernetes.io/projected/173e3771-e034-4856-ba30-9eecf64a83c6-kube-api-access-br8th\") pod \"community-operators-5kk9g\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.916359 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-utilities\") pod \"community-operators-5kk9g\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.916903 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-utilities\") pod \"community-operators-5kk9g\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.916974 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-catalog-content\") pod \"community-operators-5kk9g\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:47 crc kubenswrapper[4810]: I0930 09:00:47.939503 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br8th\" (UniqueName: \"kubernetes.io/projected/173e3771-e034-4856-ba30-9eecf64a83c6-kube-api-access-br8th\") pod \"community-operators-5kk9g\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:48 crc kubenswrapper[4810]: I0930 09:00:48.014899 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:48 crc kubenswrapper[4810]: I0930 09:00:48.526618 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5kk9g"] Sep 30 09:00:48 crc kubenswrapper[4810]: W0930 09:00:48.539585 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod173e3771_e034_4856_ba30_9eecf64a83c6.slice/crio-f5199446d535200a52d18ae79a8ce6919d6c442f713c974ee22e6345bb1377e2 WatchSource:0}: Error finding container f5199446d535200a52d18ae79a8ce6919d6c442f713c974ee22e6345bb1377e2: Status 404 returned error can't find the container with id f5199446d535200a52d18ae79a8ce6919d6c442f713c974ee22e6345bb1377e2 Sep 30 09:00:49 crc kubenswrapper[4810]: I0930 09:00:49.023131 4810 generic.go:334] "Generic (PLEG): container finished" podID="173e3771-e034-4856-ba30-9eecf64a83c6" containerID="b27bf2299eb3811384a8e14810dd068d6c8a370b68939fc3dbeecbdb86f96b9d" exitCode=0 Sep 30 09:00:49 crc kubenswrapper[4810]: I0930 09:00:49.023304 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kk9g" event={"ID":"173e3771-e034-4856-ba30-9eecf64a83c6","Type":"ContainerDied","Data":"b27bf2299eb3811384a8e14810dd068d6c8a370b68939fc3dbeecbdb86f96b9d"} Sep 30 09:00:49 crc kubenswrapper[4810]: I0930 09:00:49.023523 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kk9g" event={"ID":"173e3771-e034-4856-ba30-9eecf64a83c6","Type":"ContainerStarted","Data":"f5199446d535200a52d18ae79a8ce6919d6c442f713c974ee22e6345bb1377e2"} Sep 30 09:00:50 crc kubenswrapper[4810]: I0930 09:00:50.034081 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kk9g" event={"ID":"173e3771-e034-4856-ba30-9eecf64a83c6","Type":"ContainerStarted","Data":"55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa"} Sep 30 09:00:52 crc kubenswrapper[4810]: I0930 09:00:52.058458 4810 generic.go:334] "Generic (PLEG): container finished" podID="173e3771-e034-4856-ba30-9eecf64a83c6" containerID="55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa" exitCode=0 Sep 30 09:00:52 crc kubenswrapper[4810]: I0930 09:00:52.058510 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kk9g" event={"ID":"173e3771-e034-4856-ba30-9eecf64a83c6","Type":"ContainerDied","Data":"55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa"} Sep 30 09:00:53 crc kubenswrapper[4810]: I0930 09:00:53.084697 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kk9g" event={"ID":"173e3771-e034-4856-ba30-9eecf64a83c6","Type":"ContainerStarted","Data":"7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606"} Sep 30 09:00:53 crc kubenswrapper[4810]: I0930 09:00:53.112787 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5kk9g" podStartSLOduration=2.511136078 podStartE2EDuration="6.112769182s" podCreationTimestamp="2025-09-30 09:00:47 +0000 UTC" firstStartedPulling="2025-09-30 09:00:49.025945304 +0000 UTC m=+3472.478144571" lastFinishedPulling="2025-09-30 09:00:52.627578368 +0000 UTC m=+3476.079777675" observedRunningTime="2025-09-30 09:00:53.110134325 +0000 UTC m=+3476.562333642" watchObservedRunningTime="2025-09-30 09:00:53.112769182 +0000 UTC m=+3476.564968459" Sep 30 09:00:58 crc kubenswrapper[4810]: I0930 09:00:58.015814 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:58 crc kubenswrapper[4810]: I0930 09:00:58.016409 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:58 crc kubenswrapper[4810]: I0930 09:00:58.098108 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:58 crc kubenswrapper[4810]: I0930 09:00:58.191762 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:00:58 crc kubenswrapper[4810]: I0930 09:00:58.342128 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5kk9g"] Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.167722 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29320381-wgv24"] Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.169588 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.178202 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29320381-wgv24"] Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.215648 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5kk9g" podUID="173e3771-e034-4856-ba30-9eecf64a83c6" containerName="registry-server" containerID="cri-o://7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606" gracePeriod=2 Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.301219 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-fernet-keys\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.301447 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-config-data\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.301475 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5f4m\" (UniqueName: \"kubernetes.io/projected/dd37005c-311f-4aef-9f91-bc01d5d62249-kube-api-access-s5f4m\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.301655 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-combined-ca-bundle\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.403817 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-config-data\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.404672 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5f4m\" (UniqueName: \"kubernetes.io/projected/dd37005c-311f-4aef-9f91-bc01d5d62249-kube-api-access-s5f4m\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.404770 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-combined-ca-bundle\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.404820 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-fernet-keys\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.411465 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-config-data\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.412430 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-combined-ca-bundle\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.412732 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-fernet-keys\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.431165 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5f4m\" (UniqueName: \"kubernetes.io/projected/dd37005c-311f-4aef-9f91-bc01d5d62249-kube-api-access-s5f4m\") pod \"keystone-cron-29320381-wgv24\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.519084 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.649878 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.710285 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-catalog-content\") pod \"173e3771-e034-4856-ba30-9eecf64a83c6\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.710439 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br8th\" (UniqueName: \"kubernetes.io/projected/173e3771-e034-4856-ba30-9eecf64a83c6-kube-api-access-br8th\") pod \"173e3771-e034-4856-ba30-9eecf64a83c6\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.710500 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-utilities\") pod \"173e3771-e034-4856-ba30-9eecf64a83c6\" (UID: \"173e3771-e034-4856-ba30-9eecf64a83c6\") " Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.711882 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-utilities" (OuterVolumeSpecName: "utilities") pod "173e3771-e034-4856-ba30-9eecf64a83c6" (UID: "173e3771-e034-4856-ba30-9eecf64a83c6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.717229 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/173e3771-e034-4856-ba30-9eecf64a83c6-kube-api-access-br8th" (OuterVolumeSpecName: "kube-api-access-br8th") pod "173e3771-e034-4856-ba30-9eecf64a83c6" (UID: "173e3771-e034-4856-ba30-9eecf64a83c6"). InnerVolumeSpecName "kube-api-access-br8th". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.759992 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "173e3771-e034-4856-ba30-9eecf64a83c6" (UID: "173e3771-e034-4856-ba30-9eecf64a83c6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.812965 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.812996 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br8th\" (UniqueName: \"kubernetes.io/projected/173e3771-e034-4856-ba30-9eecf64a83c6-kube-api-access-br8th\") on node \"crc\" DevicePath \"\"" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.813005 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173e3771-e034-4856-ba30-9eecf64a83c6-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:01:00 crc kubenswrapper[4810]: I0930 09:01:00.986681 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29320381-wgv24"] Sep 30 09:01:00 crc kubenswrapper[4810]: W0930 09:01:00.992674 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd37005c_311f_4aef_9f91_bc01d5d62249.slice/crio-603bc7a311f6a795a32e2f65efc1481e23236bca1542fd0ed7a0bdb3171b8923 WatchSource:0}: Error finding container 603bc7a311f6a795a32e2f65efc1481e23236bca1542fd0ed7a0bdb3171b8923: Status 404 returned error can't find the container with id 603bc7a311f6a795a32e2f65efc1481e23236bca1542fd0ed7a0bdb3171b8923 Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.228921 4810 generic.go:334] "Generic (PLEG): container finished" podID="173e3771-e034-4856-ba30-9eecf64a83c6" containerID="7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606" exitCode=0 Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.229017 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kk9g" Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.229029 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kk9g" event={"ID":"173e3771-e034-4856-ba30-9eecf64a83c6","Type":"ContainerDied","Data":"7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606"} Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.230238 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kk9g" event={"ID":"173e3771-e034-4856-ba30-9eecf64a83c6","Type":"ContainerDied","Data":"f5199446d535200a52d18ae79a8ce6919d6c442f713c974ee22e6345bb1377e2"} Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.230267 4810 scope.go:117] "RemoveContainer" containerID="7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606" Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.231242 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320381-wgv24" event={"ID":"dd37005c-311f-4aef-9f91-bc01d5d62249","Type":"ContainerStarted","Data":"e8f71d6e7417ebd6a503ca1105ffff6d9a05756141020a41ce75a32371fe2d71"} Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.231267 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320381-wgv24" event={"ID":"dd37005c-311f-4aef-9f91-bc01d5d62249","Type":"ContainerStarted","Data":"603bc7a311f6a795a32e2f65efc1481e23236bca1542fd0ed7a0bdb3171b8923"} Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.261532 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29320381-wgv24" podStartSLOduration=1.261501174 podStartE2EDuration="1.261501174s" podCreationTimestamp="2025-09-30 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 09:01:01.248583958 +0000 UTC m=+3484.700783225" watchObservedRunningTime="2025-09-30 09:01:01.261501174 +0000 UTC m=+3484.713700451" Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.269661 4810 scope.go:117] "RemoveContainer" containerID="55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa" Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.279365 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5kk9g"] Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.289502 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5kk9g"] Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.297021 4810 scope.go:117] "RemoveContainer" containerID="b27bf2299eb3811384a8e14810dd068d6c8a370b68939fc3dbeecbdb86f96b9d" Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.320950 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="173e3771-e034-4856-ba30-9eecf64a83c6" path="/var/lib/kubelet/pods/173e3771-e034-4856-ba30-9eecf64a83c6/volumes" Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.337041 4810 scope.go:117] "RemoveContainer" containerID="7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606" Sep 30 09:01:01 crc kubenswrapper[4810]: E0930 09:01:01.337620 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606\": container with ID starting with 7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606 not found: ID does not exist" containerID="7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606" Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.337668 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606"} err="failed to get container status \"7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606\": rpc error: code = NotFound desc = could not find container \"7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606\": container with ID starting with 7b95f50f464cc5d19dfeaab6a7f25450db1256b1a3de8d0492df8e9e0f98a606 not found: ID does not exist" Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.337699 4810 scope.go:117] "RemoveContainer" containerID="55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa" Sep 30 09:01:01 crc kubenswrapper[4810]: E0930 09:01:01.338092 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa\": container with ID starting with 55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa not found: ID does not exist" containerID="55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa" Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.338124 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa"} err="failed to get container status \"55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa\": rpc error: code = NotFound desc = could not find container \"55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa\": container with ID starting with 55cf1e3669f1841aa44de76d7a6ee9192c4c4d45a78b5e137ddc238cd189ceaa not found: ID does not exist" Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.338143 4810 scope.go:117] "RemoveContainer" containerID="b27bf2299eb3811384a8e14810dd068d6c8a370b68939fc3dbeecbdb86f96b9d" Sep 30 09:01:01 crc kubenswrapper[4810]: E0930 09:01:01.338430 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b27bf2299eb3811384a8e14810dd068d6c8a370b68939fc3dbeecbdb86f96b9d\": container with ID starting with b27bf2299eb3811384a8e14810dd068d6c8a370b68939fc3dbeecbdb86f96b9d not found: ID does not exist" containerID="b27bf2299eb3811384a8e14810dd068d6c8a370b68939fc3dbeecbdb86f96b9d" Sep 30 09:01:01 crc kubenswrapper[4810]: I0930 09:01:01.338463 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b27bf2299eb3811384a8e14810dd068d6c8a370b68939fc3dbeecbdb86f96b9d"} err="failed to get container status \"b27bf2299eb3811384a8e14810dd068d6c8a370b68939fc3dbeecbdb86f96b9d\": rpc error: code = NotFound desc = could not find container \"b27bf2299eb3811384a8e14810dd068d6c8a370b68939fc3dbeecbdb86f96b9d\": container with ID starting with b27bf2299eb3811384a8e14810dd068d6c8a370b68939fc3dbeecbdb86f96b9d not found: ID does not exist" Sep 30 09:01:05 crc kubenswrapper[4810]: I0930 09:01:05.285371 4810 generic.go:334] "Generic (PLEG): container finished" podID="dd37005c-311f-4aef-9f91-bc01d5d62249" containerID="e8f71d6e7417ebd6a503ca1105ffff6d9a05756141020a41ce75a32371fe2d71" exitCode=0 Sep 30 09:01:05 crc kubenswrapper[4810]: I0930 09:01:05.285498 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320381-wgv24" event={"ID":"dd37005c-311f-4aef-9f91-bc01d5d62249","Type":"ContainerDied","Data":"e8f71d6e7417ebd6a503ca1105ffff6d9a05756141020a41ce75a32371fe2d71"} Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.612229 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.749909 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-combined-ca-bundle\") pod \"dd37005c-311f-4aef-9f91-bc01d5d62249\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.750120 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5f4m\" (UniqueName: \"kubernetes.io/projected/dd37005c-311f-4aef-9f91-bc01d5d62249-kube-api-access-s5f4m\") pod \"dd37005c-311f-4aef-9f91-bc01d5d62249\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.750150 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-fernet-keys\") pod \"dd37005c-311f-4aef-9f91-bc01d5d62249\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.750178 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-config-data\") pod \"dd37005c-311f-4aef-9f91-bc01d5d62249\" (UID: \"dd37005c-311f-4aef-9f91-bc01d5d62249\") " Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.755370 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "dd37005c-311f-4aef-9f91-bc01d5d62249" (UID: "dd37005c-311f-4aef-9f91-bc01d5d62249"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.756723 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd37005c-311f-4aef-9f91-bc01d5d62249-kube-api-access-s5f4m" (OuterVolumeSpecName: "kube-api-access-s5f4m") pod "dd37005c-311f-4aef-9f91-bc01d5d62249" (UID: "dd37005c-311f-4aef-9f91-bc01d5d62249"). InnerVolumeSpecName "kube-api-access-s5f4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.782327 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd37005c-311f-4aef-9f91-bc01d5d62249" (UID: "dd37005c-311f-4aef-9f91-bc01d5d62249"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.804707 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-config-data" (OuterVolumeSpecName: "config-data") pod "dd37005c-311f-4aef-9f91-bc01d5d62249" (UID: "dd37005c-311f-4aef-9f91-bc01d5d62249"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.852585 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.852634 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5f4m\" (UniqueName: \"kubernetes.io/projected/dd37005c-311f-4aef-9f91-bc01d5d62249-kube-api-access-s5f4m\") on node \"crc\" DevicePath \"\"" Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.852645 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 09:01:06 crc kubenswrapper[4810]: I0930 09:01:06.852653 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd37005c-311f-4aef-9f91-bc01d5d62249-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 09:01:07 crc kubenswrapper[4810]: I0930 09:01:07.327482 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320381-wgv24" Sep 30 09:01:07 crc kubenswrapper[4810]: I0930 09:01:07.332661 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320381-wgv24" event={"ID":"dd37005c-311f-4aef-9f91-bc01d5d62249","Type":"ContainerDied","Data":"603bc7a311f6a795a32e2f65efc1481e23236bca1542fd0ed7a0bdb3171b8923"} Sep 30 09:01:07 crc kubenswrapper[4810]: I0930 09:01:07.332720 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="603bc7a311f6a795a32e2f65efc1481e23236bca1542fd0ed7a0bdb3171b8923" Sep 30 09:02:45 crc kubenswrapper[4810]: I0930 09:02:45.911762 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:02:45 crc kubenswrapper[4810]: I0930 09:02:45.913608 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:03:15 crc kubenswrapper[4810]: I0930 09:03:15.912015 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:03:15 crc kubenswrapper[4810]: I0930 09:03:15.912655 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:03:45 crc kubenswrapper[4810]: I0930 09:03:45.912242 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:03:45 crc kubenswrapper[4810]: I0930 09:03:45.912854 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:03:45 crc kubenswrapper[4810]: I0930 09:03:45.912945 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 09:03:45 crc kubenswrapper[4810]: I0930 09:03:45.914534 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b04c10958aa5c72edbc42e5ad3647261d4e99fe293d352aeae3dd3f087658bc1"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 09:03:45 crc kubenswrapper[4810]: I0930 09:03:45.914671 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://b04c10958aa5c72edbc42e5ad3647261d4e99fe293d352aeae3dd3f087658bc1" gracePeriod=600 Sep 30 09:03:46 crc kubenswrapper[4810]: I0930 09:03:46.189770 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="b04c10958aa5c72edbc42e5ad3647261d4e99fe293d352aeae3dd3f087658bc1" exitCode=0 Sep 30 09:03:46 crc kubenswrapper[4810]: I0930 09:03:46.189829 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"b04c10958aa5c72edbc42e5ad3647261d4e99fe293d352aeae3dd3f087658bc1"} Sep 30 09:03:46 crc kubenswrapper[4810]: I0930 09:03:46.190209 4810 scope.go:117] "RemoveContainer" containerID="57727b5f11e13177b58e4d1b0d9ad30fe9ea4e3619f15ba1d5fdaea1aecd86a5" Sep 30 09:03:47 crc kubenswrapper[4810]: I0930 09:03:47.205370 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f"} Sep 30 09:06:15 crc kubenswrapper[4810]: I0930 09:06:15.912185 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:06:15 crc kubenswrapper[4810]: I0930 09:06:15.912819 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:06:45 crc kubenswrapper[4810]: I0930 09:06:45.911639 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:06:45 crc kubenswrapper[4810]: I0930 09:06:45.912455 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:07:15 crc kubenswrapper[4810]: I0930 09:07:15.911480 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:07:15 crc kubenswrapper[4810]: I0930 09:07:15.912244 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:07:15 crc kubenswrapper[4810]: I0930 09:07:15.912351 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 09:07:15 crc kubenswrapper[4810]: I0930 09:07:15.913555 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 09:07:15 crc kubenswrapper[4810]: I0930 09:07:15.913694 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" gracePeriod=600 Sep 30 09:07:16 crc kubenswrapper[4810]: E0930 09:07:16.040330 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:07:16 crc kubenswrapper[4810]: I0930 09:07:16.539001 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" exitCode=0 Sep 30 09:07:16 crc kubenswrapper[4810]: I0930 09:07:16.539079 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f"} Sep 30 09:07:16 crc kubenswrapper[4810]: I0930 09:07:16.539423 4810 scope.go:117] "RemoveContainer" containerID="b04c10958aa5c72edbc42e5ad3647261d4e99fe293d352aeae3dd3f087658bc1" Sep 30 09:07:16 crc kubenswrapper[4810]: I0930 09:07:16.540488 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:07:16 crc kubenswrapper[4810]: E0930 09:07:16.541049 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.589659 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w5gz9"] Sep 30 09:07:22 crc kubenswrapper[4810]: E0930 09:07:22.590923 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173e3771-e034-4856-ba30-9eecf64a83c6" containerName="registry-server" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.590948 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="173e3771-e034-4856-ba30-9eecf64a83c6" containerName="registry-server" Sep 30 09:07:22 crc kubenswrapper[4810]: E0930 09:07:22.591021 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd37005c-311f-4aef-9f91-bc01d5d62249" containerName="keystone-cron" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.591035 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd37005c-311f-4aef-9f91-bc01d5d62249" containerName="keystone-cron" Sep 30 09:07:22 crc kubenswrapper[4810]: E0930 09:07:22.591074 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173e3771-e034-4856-ba30-9eecf64a83c6" containerName="extract-content" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.591087 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="173e3771-e034-4856-ba30-9eecf64a83c6" containerName="extract-content" Sep 30 09:07:22 crc kubenswrapper[4810]: E0930 09:07:22.591117 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173e3771-e034-4856-ba30-9eecf64a83c6" containerName="extract-utilities" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.591137 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="173e3771-e034-4856-ba30-9eecf64a83c6" containerName="extract-utilities" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.591566 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd37005c-311f-4aef-9f91-bc01d5d62249" containerName="keystone-cron" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.591616 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="173e3771-e034-4856-ba30-9eecf64a83c6" containerName="registry-server" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.598427 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.626281 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5gz9"] Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.706140 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r5nd\" (UniqueName: \"kubernetes.io/projected/33c345df-495c-4f17-885e-377ecd85b599-kube-api-access-9r5nd\") pod \"redhat-marketplace-w5gz9\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.706449 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-catalog-content\") pod \"redhat-marketplace-w5gz9\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.706535 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-utilities\") pod \"redhat-marketplace-w5gz9\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.808615 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-utilities\") pod \"redhat-marketplace-w5gz9\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.808797 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r5nd\" (UniqueName: \"kubernetes.io/projected/33c345df-495c-4f17-885e-377ecd85b599-kube-api-access-9r5nd\") pod \"redhat-marketplace-w5gz9\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.808825 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-catalog-content\") pod \"redhat-marketplace-w5gz9\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.809170 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-utilities\") pod \"redhat-marketplace-w5gz9\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.809407 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-catalog-content\") pod \"redhat-marketplace-w5gz9\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.839100 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r5nd\" (UniqueName: \"kubernetes.io/projected/33c345df-495c-4f17-885e-377ecd85b599-kube-api-access-9r5nd\") pod \"redhat-marketplace-w5gz9\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:22 crc kubenswrapper[4810]: I0930 09:07:22.943857 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:23 crc kubenswrapper[4810]: I0930 09:07:23.461840 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5gz9"] Sep 30 09:07:23 crc kubenswrapper[4810]: I0930 09:07:23.634718 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5gz9" event={"ID":"33c345df-495c-4f17-885e-377ecd85b599","Type":"ContainerStarted","Data":"ece54bc21cd19aac84208c75faafb7f08168077b1db59daae36e4e1c7870394a"} Sep 30 09:07:24 crc kubenswrapper[4810]: I0930 09:07:24.657106 4810 generic.go:334] "Generic (PLEG): container finished" podID="33c345df-495c-4f17-885e-377ecd85b599" containerID="3e664260a0671e9b7f2a4a01f742eecbcc0332afd5eb50b5415f6bc87a027b8d" exitCode=0 Sep 30 09:07:24 crc kubenswrapper[4810]: I0930 09:07:24.657562 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5gz9" event={"ID":"33c345df-495c-4f17-885e-377ecd85b599","Type":"ContainerDied","Data":"3e664260a0671e9b7f2a4a01f742eecbcc0332afd5eb50b5415f6bc87a027b8d"} Sep 30 09:07:24 crc kubenswrapper[4810]: I0930 09:07:24.661452 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 09:07:25 crc kubenswrapper[4810]: I0930 09:07:25.675810 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5gz9" event={"ID":"33c345df-495c-4f17-885e-377ecd85b599","Type":"ContainerStarted","Data":"561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947"} Sep 30 09:07:26 crc kubenswrapper[4810]: I0930 09:07:26.686938 4810 generic.go:334] "Generic (PLEG): container finished" podID="33c345df-495c-4f17-885e-377ecd85b599" containerID="561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947" exitCode=0 Sep 30 09:07:26 crc kubenswrapper[4810]: I0930 09:07:26.687020 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5gz9" event={"ID":"33c345df-495c-4f17-885e-377ecd85b599","Type":"ContainerDied","Data":"561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947"} Sep 30 09:07:27 crc kubenswrapper[4810]: I0930 09:07:27.699529 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5gz9" event={"ID":"33c345df-495c-4f17-885e-377ecd85b599","Type":"ContainerStarted","Data":"9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787"} Sep 30 09:07:27 crc kubenswrapper[4810]: I0930 09:07:27.723058 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w5gz9" podStartSLOduration=3.065670765 podStartE2EDuration="5.72303645s" podCreationTimestamp="2025-09-30 09:07:22 +0000 UTC" firstStartedPulling="2025-09-30 09:07:24.66054664 +0000 UTC m=+3868.112745947" lastFinishedPulling="2025-09-30 09:07:27.317912365 +0000 UTC m=+3870.770111632" observedRunningTime="2025-09-30 09:07:27.715866032 +0000 UTC m=+3871.168065309" watchObservedRunningTime="2025-09-30 09:07:27.72303645 +0000 UTC m=+3871.175235717" Sep 30 09:07:31 crc kubenswrapper[4810]: I0930 09:07:31.306905 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:07:31 crc kubenswrapper[4810]: E0930 09:07:31.309013 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:07:32 crc kubenswrapper[4810]: I0930 09:07:32.944002 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:32 crc kubenswrapper[4810]: I0930 09:07:32.944819 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:32 crc kubenswrapper[4810]: I0930 09:07:32.994384 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:33 crc kubenswrapper[4810]: I0930 09:07:33.832364 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:33 crc kubenswrapper[4810]: I0930 09:07:33.893779 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5gz9"] Sep 30 09:07:35 crc kubenswrapper[4810]: I0930 09:07:35.780355 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w5gz9" podUID="33c345df-495c-4f17-885e-377ecd85b599" containerName="registry-server" containerID="cri-o://9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787" gracePeriod=2 Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.295711 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.393816 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-utilities\") pod \"33c345df-495c-4f17-885e-377ecd85b599\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.394047 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-catalog-content\") pod \"33c345df-495c-4f17-885e-377ecd85b599\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.394125 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r5nd\" (UniqueName: \"kubernetes.io/projected/33c345df-495c-4f17-885e-377ecd85b599-kube-api-access-9r5nd\") pod \"33c345df-495c-4f17-885e-377ecd85b599\" (UID: \"33c345df-495c-4f17-885e-377ecd85b599\") " Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.395398 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-utilities" (OuterVolumeSpecName: "utilities") pod "33c345df-495c-4f17-885e-377ecd85b599" (UID: "33c345df-495c-4f17-885e-377ecd85b599"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.403851 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33c345df-495c-4f17-885e-377ecd85b599-kube-api-access-9r5nd" (OuterVolumeSpecName: "kube-api-access-9r5nd") pod "33c345df-495c-4f17-885e-377ecd85b599" (UID: "33c345df-495c-4f17-885e-377ecd85b599"). InnerVolumeSpecName "kube-api-access-9r5nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.407921 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33c345df-495c-4f17-885e-377ecd85b599" (UID: "33c345df-495c-4f17-885e-377ecd85b599"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.496237 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r5nd\" (UniqueName: \"kubernetes.io/projected/33c345df-495c-4f17-885e-377ecd85b599-kube-api-access-9r5nd\") on node \"crc\" DevicePath \"\"" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.496286 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.496299 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c345df-495c-4f17-885e-377ecd85b599-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.798787 4810 generic.go:334] "Generic (PLEG): container finished" podID="33c345df-495c-4f17-885e-377ecd85b599" containerID="9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787" exitCode=0 Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.798867 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5gz9" event={"ID":"33c345df-495c-4f17-885e-377ecd85b599","Type":"ContainerDied","Data":"9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787"} Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.798896 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5gz9" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.798930 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5gz9" event={"ID":"33c345df-495c-4f17-885e-377ecd85b599","Type":"ContainerDied","Data":"ece54bc21cd19aac84208c75faafb7f08168077b1db59daae36e4e1c7870394a"} Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.798972 4810 scope.go:117] "RemoveContainer" containerID="9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.850554 4810 scope.go:117] "RemoveContainer" containerID="561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.866292 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5gz9"] Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.877849 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5gz9"] Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.896042 4810 scope.go:117] "RemoveContainer" containerID="3e664260a0671e9b7f2a4a01f742eecbcc0332afd5eb50b5415f6bc87a027b8d" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.922735 4810 scope.go:117] "RemoveContainer" containerID="9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787" Sep 30 09:07:36 crc kubenswrapper[4810]: E0930 09:07:36.923154 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787\": container with ID starting with 9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787 not found: ID does not exist" containerID="9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.923205 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787"} err="failed to get container status \"9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787\": rpc error: code = NotFound desc = could not find container \"9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787\": container with ID starting with 9007b02c95139091c6a203df4634c2c19eb6ab754e3899faed881eb58b231787 not found: ID does not exist" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.923240 4810 scope.go:117] "RemoveContainer" containerID="561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947" Sep 30 09:07:36 crc kubenswrapper[4810]: E0930 09:07:36.923578 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947\": container with ID starting with 561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947 not found: ID does not exist" containerID="561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.923629 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947"} err="failed to get container status \"561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947\": rpc error: code = NotFound desc = could not find container \"561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947\": container with ID starting with 561aef9c971a5af843766970ba889f54df8842b04546ebfc540658f6daaa3947 not found: ID does not exist" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.923654 4810 scope.go:117] "RemoveContainer" containerID="3e664260a0671e9b7f2a4a01f742eecbcc0332afd5eb50b5415f6bc87a027b8d" Sep 30 09:07:36 crc kubenswrapper[4810]: E0930 09:07:36.923918 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e664260a0671e9b7f2a4a01f742eecbcc0332afd5eb50b5415f6bc87a027b8d\": container with ID starting with 3e664260a0671e9b7f2a4a01f742eecbcc0332afd5eb50b5415f6bc87a027b8d not found: ID does not exist" containerID="3e664260a0671e9b7f2a4a01f742eecbcc0332afd5eb50b5415f6bc87a027b8d" Sep 30 09:07:36 crc kubenswrapper[4810]: I0930 09:07:36.923963 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e664260a0671e9b7f2a4a01f742eecbcc0332afd5eb50b5415f6bc87a027b8d"} err="failed to get container status \"3e664260a0671e9b7f2a4a01f742eecbcc0332afd5eb50b5415f6bc87a027b8d\": rpc error: code = NotFound desc = could not find container \"3e664260a0671e9b7f2a4a01f742eecbcc0332afd5eb50b5415f6bc87a027b8d\": container with ID starting with 3e664260a0671e9b7f2a4a01f742eecbcc0332afd5eb50b5415f6bc87a027b8d not found: ID does not exist" Sep 30 09:07:37 crc kubenswrapper[4810]: I0930 09:07:37.327838 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33c345df-495c-4f17-885e-377ecd85b599" path="/var/lib/kubelet/pods/33c345df-495c-4f17-885e-377ecd85b599/volumes" Sep 30 09:07:46 crc kubenswrapper[4810]: I0930 09:07:46.307128 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:07:46 crc kubenswrapper[4810]: E0930 09:07:46.308227 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:08:00 crc kubenswrapper[4810]: I0930 09:08:00.308034 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:08:00 crc kubenswrapper[4810]: E0930 09:08:00.309351 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:08:13 crc kubenswrapper[4810]: I0930 09:08:13.307226 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:08:13 crc kubenswrapper[4810]: E0930 09:08:13.308080 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:08:24 crc kubenswrapper[4810]: I0930 09:08:24.307258 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:08:24 crc kubenswrapper[4810]: E0930 09:08:24.308667 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:08:39 crc kubenswrapper[4810]: I0930 09:08:39.307617 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:08:39 crc kubenswrapper[4810]: E0930 09:08:39.308274 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.740473 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2272n"] Sep 30 09:08:41 crc kubenswrapper[4810]: E0930 09:08:41.741330 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33c345df-495c-4f17-885e-377ecd85b599" containerName="extract-utilities" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.741347 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="33c345df-495c-4f17-885e-377ecd85b599" containerName="extract-utilities" Sep 30 09:08:41 crc kubenswrapper[4810]: E0930 09:08:41.741395 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33c345df-495c-4f17-885e-377ecd85b599" containerName="registry-server" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.741404 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="33c345df-495c-4f17-885e-377ecd85b599" containerName="registry-server" Sep 30 09:08:41 crc kubenswrapper[4810]: E0930 09:08:41.741420 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33c345df-495c-4f17-885e-377ecd85b599" containerName="extract-content" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.741430 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="33c345df-495c-4f17-885e-377ecd85b599" containerName="extract-content" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.741683 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="33c345df-495c-4f17-885e-377ecd85b599" containerName="registry-server" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.743476 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.764785 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2272n"] Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.844846 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-utilities\") pod \"certified-operators-2272n\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.844900 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-catalog-content\") pod \"certified-operators-2272n\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.844920 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8h5n\" (UniqueName: \"kubernetes.io/projected/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-kube-api-access-v8h5n\") pod \"certified-operators-2272n\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.946648 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-utilities\") pod \"certified-operators-2272n\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.946708 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-catalog-content\") pod \"certified-operators-2272n\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.946735 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8h5n\" (UniqueName: \"kubernetes.io/projected/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-kube-api-access-v8h5n\") pod \"certified-operators-2272n\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.947160 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-catalog-content\") pod \"certified-operators-2272n\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.947228 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-utilities\") pod \"certified-operators-2272n\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:41 crc kubenswrapper[4810]: I0930 09:08:41.965771 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8h5n\" (UniqueName: \"kubernetes.io/projected/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-kube-api-access-v8h5n\") pod \"certified-operators-2272n\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:42 crc kubenswrapper[4810]: I0930 09:08:42.076423 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:42 crc kubenswrapper[4810]: I0930 09:08:42.589115 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2272n"] Sep 30 09:08:43 crc kubenswrapper[4810]: I0930 09:08:43.571789 4810 generic.go:334] "Generic (PLEG): container finished" podID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" containerID="72f3ecb1f173b409d561d4600033cc578422f5da5ae47261d8bc28a6c407cad9" exitCode=0 Sep 30 09:08:43 crc kubenswrapper[4810]: I0930 09:08:43.571924 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2272n" event={"ID":"79e56903-ffc4-4e23-b1bf-1a3a9dabe739","Type":"ContainerDied","Data":"72f3ecb1f173b409d561d4600033cc578422f5da5ae47261d8bc28a6c407cad9"} Sep 30 09:08:43 crc kubenswrapper[4810]: I0930 09:08:43.572071 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2272n" event={"ID":"79e56903-ffc4-4e23-b1bf-1a3a9dabe739","Type":"ContainerStarted","Data":"042e8853d34943b49c88410c7eaaeeaacb69262b4c2969714b5302a188471045"} Sep 30 09:08:44 crc kubenswrapper[4810]: I0930 09:08:44.582635 4810 generic.go:334] "Generic (PLEG): container finished" podID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" containerID="e507a662df5b054b8fd2245360b2f0928a7d1334ea97d86425fd12a60a14b512" exitCode=0 Sep 30 09:08:44 crc kubenswrapper[4810]: I0930 09:08:44.582758 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2272n" event={"ID":"79e56903-ffc4-4e23-b1bf-1a3a9dabe739","Type":"ContainerDied","Data":"e507a662df5b054b8fd2245360b2f0928a7d1334ea97d86425fd12a60a14b512"} Sep 30 09:08:45 crc kubenswrapper[4810]: I0930 09:08:45.595045 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2272n" event={"ID":"79e56903-ffc4-4e23-b1bf-1a3a9dabe739","Type":"ContainerStarted","Data":"d5a115b07989ccc126b6d6692af6d30c6177610fe11065984fa52c8d55f42476"} Sep 30 09:08:45 crc kubenswrapper[4810]: I0930 09:08:45.630609 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2272n" podStartSLOduration=2.94142415 podStartE2EDuration="4.630588765s" podCreationTimestamp="2025-09-30 09:08:41 +0000 UTC" firstStartedPulling="2025-09-30 09:08:43.574519852 +0000 UTC m=+3947.026719159" lastFinishedPulling="2025-09-30 09:08:45.263684517 +0000 UTC m=+3948.715883774" observedRunningTime="2025-09-30 09:08:45.621046839 +0000 UTC m=+3949.073246116" watchObservedRunningTime="2025-09-30 09:08:45.630588765 +0000 UTC m=+3949.082788032" Sep 30 09:08:52 crc kubenswrapper[4810]: I0930 09:08:52.077406 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:52 crc kubenswrapper[4810]: I0930 09:08:52.078094 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:52 crc kubenswrapper[4810]: I0930 09:08:52.126363 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:52 crc kubenswrapper[4810]: I0930 09:08:52.307072 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:08:52 crc kubenswrapper[4810]: E0930 09:08:52.307557 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:08:52 crc kubenswrapper[4810]: I0930 09:08:52.740929 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:52 crc kubenswrapper[4810]: I0930 09:08:52.795542 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2272n"] Sep 30 09:08:54 crc kubenswrapper[4810]: I0930 09:08:54.719155 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2272n" podUID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" containerName="registry-server" containerID="cri-o://d5a115b07989ccc126b6d6692af6d30c6177610fe11065984fa52c8d55f42476" gracePeriod=2 Sep 30 09:08:55 crc kubenswrapper[4810]: I0930 09:08:55.730067 4810 generic.go:334] "Generic (PLEG): container finished" podID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" containerID="d5a115b07989ccc126b6d6692af6d30c6177610fe11065984fa52c8d55f42476" exitCode=0 Sep 30 09:08:55 crc kubenswrapper[4810]: I0930 09:08:55.730115 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2272n" event={"ID":"79e56903-ffc4-4e23-b1bf-1a3a9dabe739","Type":"ContainerDied","Data":"d5a115b07989ccc126b6d6692af6d30c6177610fe11065984fa52c8d55f42476"} Sep 30 09:08:55 crc kubenswrapper[4810]: I0930 09:08:55.730382 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2272n" event={"ID":"79e56903-ffc4-4e23-b1bf-1a3a9dabe739","Type":"ContainerDied","Data":"042e8853d34943b49c88410c7eaaeeaacb69262b4c2969714b5302a188471045"} Sep 30 09:08:55 crc kubenswrapper[4810]: I0930 09:08:55.730400 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="042e8853d34943b49c88410c7eaaeeaacb69262b4c2969714b5302a188471045" Sep 30 09:08:55 crc kubenswrapper[4810]: I0930 09:08:55.822512 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:55 crc kubenswrapper[4810]: I0930 09:08:55.947933 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-utilities\") pod \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " Sep 30 09:08:55 crc kubenswrapper[4810]: I0930 09:08:55.948090 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-catalog-content\") pod \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " Sep 30 09:08:55 crc kubenswrapper[4810]: I0930 09:08:55.948141 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8h5n\" (UniqueName: \"kubernetes.io/projected/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-kube-api-access-v8h5n\") pod \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\" (UID: \"79e56903-ffc4-4e23-b1bf-1a3a9dabe739\") " Sep 30 09:08:55 crc kubenswrapper[4810]: I0930 09:08:55.949650 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-utilities" (OuterVolumeSpecName: "utilities") pod "79e56903-ffc4-4e23-b1bf-1a3a9dabe739" (UID: "79e56903-ffc4-4e23-b1bf-1a3a9dabe739"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:08:55 crc kubenswrapper[4810]: I0930 09:08:55.955511 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-kube-api-access-v8h5n" (OuterVolumeSpecName: "kube-api-access-v8h5n") pod "79e56903-ffc4-4e23-b1bf-1a3a9dabe739" (UID: "79e56903-ffc4-4e23-b1bf-1a3a9dabe739"). InnerVolumeSpecName "kube-api-access-v8h5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:08:55 crc kubenswrapper[4810]: I0930 09:08:55.995588 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79e56903-ffc4-4e23-b1bf-1a3a9dabe739" (UID: "79e56903-ffc4-4e23-b1bf-1a3a9dabe739"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:08:56 crc kubenswrapper[4810]: I0930 09:08:56.050443 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:08:56 crc kubenswrapper[4810]: I0930 09:08:56.050481 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:08:56 crc kubenswrapper[4810]: I0930 09:08:56.050496 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8h5n\" (UniqueName: \"kubernetes.io/projected/79e56903-ffc4-4e23-b1bf-1a3a9dabe739-kube-api-access-v8h5n\") on node \"crc\" DevicePath \"\"" Sep 30 09:08:56 crc kubenswrapper[4810]: I0930 09:08:56.741769 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2272n" Sep 30 09:08:56 crc kubenswrapper[4810]: I0930 09:08:56.795274 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2272n"] Sep 30 09:08:56 crc kubenswrapper[4810]: I0930 09:08:56.805914 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2272n"] Sep 30 09:08:57 crc kubenswrapper[4810]: I0930 09:08:57.321109 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" path="/var/lib/kubelet/pods/79e56903-ffc4-4e23-b1bf-1a3a9dabe739/volumes" Sep 30 09:09:05 crc kubenswrapper[4810]: I0930 09:09:05.307036 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:09:05 crc kubenswrapper[4810]: E0930 09:09:05.308005 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:09:16 crc kubenswrapper[4810]: I0930 09:09:16.306468 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:09:16 crc kubenswrapper[4810]: E0930 09:09:16.307192 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.189737 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-42f5n"] Sep 30 09:09:27 crc kubenswrapper[4810]: E0930 09:09:27.191589 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" containerName="extract-utilities" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.191764 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" containerName="extract-utilities" Sep 30 09:09:27 crc kubenswrapper[4810]: E0930 09:09:27.191791 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" containerName="registry-server" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.191804 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" containerName="registry-server" Sep 30 09:09:27 crc kubenswrapper[4810]: E0930 09:09:27.191828 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" containerName="extract-content" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.191843 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" containerName="extract-content" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.192241 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="79e56903-ffc4-4e23-b1bf-1a3a9dabe739" containerName="registry-server" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.197001 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.198396 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-42f5n"] Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.363697 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-utilities\") pod \"redhat-operators-42f5n\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.363735 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxk6k\" (UniqueName: \"kubernetes.io/projected/43024ac6-515b-4141-8e8c-b7271dfbff18-kube-api-access-kxk6k\") pod \"redhat-operators-42f5n\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.363839 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-catalog-content\") pod \"redhat-operators-42f5n\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.465852 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-catalog-content\") pod \"redhat-operators-42f5n\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.466061 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxk6k\" (UniqueName: \"kubernetes.io/projected/43024ac6-515b-4141-8e8c-b7271dfbff18-kube-api-access-kxk6k\") pod \"redhat-operators-42f5n\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.466084 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-utilities\") pod \"redhat-operators-42f5n\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.466585 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-utilities\") pod \"redhat-operators-42f5n\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.466669 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-catalog-content\") pod \"redhat-operators-42f5n\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.492228 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxk6k\" (UniqueName: \"kubernetes.io/projected/43024ac6-515b-4141-8e8c-b7271dfbff18-kube-api-access-kxk6k\") pod \"redhat-operators-42f5n\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.529533 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:27 crc kubenswrapper[4810]: I0930 09:09:27.994564 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-42f5n"] Sep 30 09:09:28 crc kubenswrapper[4810]: I0930 09:09:28.162503 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42f5n" event={"ID":"43024ac6-515b-4141-8e8c-b7271dfbff18","Type":"ContainerStarted","Data":"c269cbf62d905cd338ca02b47fed7ea8191c15c908457bf60f0850e4690a6a85"} Sep 30 09:09:29 crc kubenswrapper[4810]: I0930 09:09:29.180346 4810 generic.go:334] "Generic (PLEG): container finished" podID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerID="9e05c2cd6ffab8e52cc772211c8c1394627485b0ceac90ecda74794d543956c9" exitCode=0 Sep 30 09:09:29 crc kubenswrapper[4810]: I0930 09:09:29.180709 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42f5n" event={"ID":"43024ac6-515b-4141-8e8c-b7271dfbff18","Type":"ContainerDied","Data":"9e05c2cd6ffab8e52cc772211c8c1394627485b0ceac90ecda74794d543956c9"} Sep 30 09:09:29 crc kubenswrapper[4810]: I0930 09:09:29.306682 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:09:29 crc kubenswrapper[4810]: E0930 09:09:29.307079 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:09:31 crc kubenswrapper[4810]: I0930 09:09:31.211718 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42f5n" event={"ID":"43024ac6-515b-4141-8e8c-b7271dfbff18","Type":"ContainerStarted","Data":"eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc"} Sep 30 09:09:33 crc kubenswrapper[4810]: I0930 09:09:33.249157 4810 generic.go:334] "Generic (PLEG): container finished" podID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerID="eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc" exitCode=0 Sep 30 09:09:33 crc kubenswrapper[4810]: I0930 09:09:33.249995 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42f5n" event={"ID":"43024ac6-515b-4141-8e8c-b7271dfbff18","Type":"ContainerDied","Data":"eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc"} Sep 30 09:09:35 crc kubenswrapper[4810]: I0930 09:09:35.269967 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42f5n" event={"ID":"43024ac6-515b-4141-8e8c-b7271dfbff18","Type":"ContainerStarted","Data":"e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac"} Sep 30 09:09:35 crc kubenswrapper[4810]: I0930 09:09:35.295331 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-42f5n" podStartSLOduration=3.40416698 podStartE2EDuration="8.295251268s" podCreationTimestamp="2025-09-30 09:09:27 +0000 UTC" firstStartedPulling="2025-09-30 09:09:29.18338182 +0000 UTC m=+3992.635581107" lastFinishedPulling="2025-09-30 09:09:34.074466118 +0000 UTC m=+3997.526665395" observedRunningTime="2025-09-30 09:09:35.286483935 +0000 UTC m=+3998.738683222" watchObservedRunningTime="2025-09-30 09:09:35.295251268 +0000 UTC m=+3998.747450535" Sep 30 09:09:37 crc kubenswrapper[4810]: I0930 09:09:37.530353 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:37 crc kubenswrapper[4810]: I0930 09:09:37.530625 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:38 crc kubenswrapper[4810]: I0930 09:09:38.573985 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-42f5n" podUID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerName="registry-server" probeResult="failure" output=< Sep 30 09:09:38 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Sep 30 09:09:38 crc kubenswrapper[4810]: > Sep 30 09:09:41 crc kubenswrapper[4810]: I0930 09:09:41.307484 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:09:41 crc kubenswrapper[4810]: E0930 09:09:41.308953 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:09:47 crc kubenswrapper[4810]: I0930 09:09:47.597045 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:47 crc kubenswrapper[4810]: I0930 09:09:47.678402 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:47 crc kubenswrapper[4810]: I0930 09:09:47.846044 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-42f5n"] Sep 30 09:09:49 crc kubenswrapper[4810]: I0930 09:09:49.415368 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-42f5n" podUID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerName="registry-server" containerID="cri-o://e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac" gracePeriod=2 Sep 30 09:09:49 crc kubenswrapper[4810]: I0930 09:09:49.936821 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:49 crc kubenswrapper[4810]: I0930 09:09:49.993397 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-catalog-content\") pod \"43024ac6-515b-4141-8e8c-b7271dfbff18\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " Sep 30 09:09:49 crc kubenswrapper[4810]: I0930 09:09:49.993468 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxk6k\" (UniqueName: \"kubernetes.io/projected/43024ac6-515b-4141-8e8c-b7271dfbff18-kube-api-access-kxk6k\") pod \"43024ac6-515b-4141-8e8c-b7271dfbff18\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " Sep 30 09:09:49 crc kubenswrapper[4810]: I0930 09:09:49.993517 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-utilities\") pod \"43024ac6-515b-4141-8e8c-b7271dfbff18\" (UID: \"43024ac6-515b-4141-8e8c-b7271dfbff18\") " Sep 30 09:09:49 crc kubenswrapper[4810]: I0930 09:09:49.995043 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-utilities" (OuterVolumeSpecName: "utilities") pod "43024ac6-515b-4141-8e8c-b7271dfbff18" (UID: "43024ac6-515b-4141-8e8c-b7271dfbff18"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.002890 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43024ac6-515b-4141-8e8c-b7271dfbff18-kube-api-access-kxk6k" (OuterVolumeSpecName: "kube-api-access-kxk6k") pod "43024ac6-515b-4141-8e8c-b7271dfbff18" (UID: "43024ac6-515b-4141-8e8c-b7271dfbff18"). InnerVolumeSpecName "kube-api-access-kxk6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.093682 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43024ac6-515b-4141-8e8c-b7271dfbff18" (UID: "43024ac6-515b-4141-8e8c-b7271dfbff18"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.096919 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.096951 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxk6k\" (UniqueName: \"kubernetes.io/projected/43024ac6-515b-4141-8e8c-b7271dfbff18-kube-api-access-kxk6k\") on node \"crc\" DevicePath \"\"" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.096966 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43024ac6-515b-4141-8e8c-b7271dfbff18-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.433804 4810 generic.go:334] "Generic (PLEG): container finished" podID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerID="e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac" exitCode=0 Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.433870 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42f5n" event={"ID":"43024ac6-515b-4141-8e8c-b7271dfbff18","Type":"ContainerDied","Data":"e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac"} Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.433954 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42f5n" event={"ID":"43024ac6-515b-4141-8e8c-b7271dfbff18","Type":"ContainerDied","Data":"c269cbf62d905cd338ca02b47fed7ea8191c15c908457bf60f0850e4690a6a85"} Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.433986 4810 scope.go:117] "RemoveContainer" containerID="e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.434014 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42f5n" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.472517 4810 scope.go:117] "RemoveContainer" containerID="eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.490776 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-42f5n"] Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.501346 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-42f5n"] Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.514169 4810 scope.go:117] "RemoveContainer" containerID="9e05c2cd6ffab8e52cc772211c8c1394627485b0ceac90ecda74794d543956c9" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.577520 4810 scope.go:117] "RemoveContainer" containerID="e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac" Sep 30 09:09:50 crc kubenswrapper[4810]: E0930 09:09:50.578457 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac\": container with ID starting with e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac not found: ID does not exist" containerID="e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.578520 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac"} err="failed to get container status \"e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac\": rpc error: code = NotFound desc = could not find container \"e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac\": container with ID starting with e0842ae6b6c2e2a1ba4ca3c448c50b4e9806100504fa851b5e0f0ac823f8aaac not found: ID does not exist" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.578573 4810 scope.go:117] "RemoveContainer" containerID="eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc" Sep 30 09:09:50 crc kubenswrapper[4810]: E0930 09:09:50.581616 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc\": container with ID starting with eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc not found: ID does not exist" containerID="eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.581723 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc"} err="failed to get container status \"eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc\": rpc error: code = NotFound desc = could not find container \"eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc\": container with ID starting with eca9fbc19738ad691bf17796ba573a970dfda9d81bd35582f1acaa1adb2483bc not found: ID does not exist" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.581769 4810 scope.go:117] "RemoveContainer" containerID="9e05c2cd6ffab8e52cc772211c8c1394627485b0ceac90ecda74794d543956c9" Sep 30 09:09:50 crc kubenswrapper[4810]: E0930 09:09:50.582335 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e05c2cd6ffab8e52cc772211c8c1394627485b0ceac90ecda74794d543956c9\": container with ID starting with 9e05c2cd6ffab8e52cc772211c8c1394627485b0ceac90ecda74794d543956c9 not found: ID does not exist" containerID="9e05c2cd6ffab8e52cc772211c8c1394627485b0ceac90ecda74794d543956c9" Sep 30 09:09:50 crc kubenswrapper[4810]: I0930 09:09:50.582377 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e05c2cd6ffab8e52cc772211c8c1394627485b0ceac90ecda74794d543956c9"} err="failed to get container status \"9e05c2cd6ffab8e52cc772211c8c1394627485b0ceac90ecda74794d543956c9\": rpc error: code = NotFound desc = could not find container \"9e05c2cd6ffab8e52cc772211c8c1394627485b0ceac90ecda74794d543956c9\": container with ID starting with 9e05c2cd6ffab8e52cc772211c8c1394627485b0ceac90ecda74794d543956c9 not found: ID does not exist" Sep 30 09:09:51 crc kubenswrapper[4810]: I0930 09:09:51.322812 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43024ac6-515b-4141-8e8c-b7271dfbff18" path="/var/lib/kubelet/pods/43024ac6-515b-4141-8e8c-b7271dfbff18/volumes" Sep 30 09:09:52 crc kubenswrapper[4810]: I0930 09:09:52.307479 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:09:52 crc kubenswrapper[4810]: E0930 09:09:52.308002 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:10:05 crc kubenswrapper[4810]: I0930 09:10:05.306879 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:10:05 crc kubenswrapper[4810]: E0930 09:10:05.307890 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:10:17 crc kubenswrapper[4810]: I0930 09:10:17.322817 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:10:17 crc kubenswrapper[4810]: E0930 09:10:17.327921 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:10:32 crc kubenswrapper[4810]: I0930 09:10:32.306728 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:10:32 crc kubenswrapper[4810]: E0930 09:10:32.307585 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:10:44 crc kubenswrapper[4810]: I0930 09:10:44.306259 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:10:44 crc kubenswrapper[4810]: E0930 09:10:44.307107 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:10:56 crc kubenswrapper[4810]: I0930 09:10:56.306776 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:10:56 crc kubenswrapper[4810]: E0930 09:10:56.307783 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:11:08 crc kubenswrapper[4810]: I0930 09:11:08.307319 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:11:08 crc kubenswrapper[4810]: E0930 09:11:08.308093 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:11:20 crc kubenswrapper[4810]: I0930 09:11:20.306350 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:11:20 crc kubenswrapper[4810]: E0930 09:11:20.307159 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:11:34 crc kubenswrapper[4810]: I0930 09:11:34.306623 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:11:34 crc kubenswrapper[4810]: E0930 09:11:34.307602 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:11:47 crc kubenswrapper[4810]: I0930 09:11:47.320893 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:11:47 crc kubenswrapper[4810]: E0930 09:11:47.322209 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:12:02 crc kubenswrapper[4810]: I0930 09:12:02.307079 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:12:02 crc kubenswrapper[4810]: E0930 09:12:02.308183 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.266893 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kr8hl"] Sep 30 09:12:07 crc kubenswrapper[4810]: E0930 09:12:07.269041 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerName="registry-server" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.269153 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerName="registry-server" Sep 30 09:12:07 crc kubenswrapper[4810]: E0930 09:12:07.269253 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerName="extract-content" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.269363 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerName="extract-content" Sep 30 09:12:07 crc kubenswrapper[4810]: E0930 09:12:07.269459 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerName="extract-utilities" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.269531 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerName="extract-utilities" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.269869 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="43024ac6-515b-4141-8e8c-b7271dfbff18" containerName="registry-server" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.271802 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.296980 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kr8hl"] Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.331213 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-catalog-content\") pod \"community-operators-kr8hl\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.331540 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-utilities\") pod \"community-operators-kr8hl\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.331767 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zxjg\" (UniqueName: \"kubernetes.io/projected/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-kube-api-access-5zxjg\") pod \"community-operators-kr8hl\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.433717 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zxjg\" (UniqueName: \"kubernetes.io/projected/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-kube-api-access-5zxjg\") pod \"community-operators-kr8hl\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.434168 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-catalog-content\") pod \"community-operators-kr8hl\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.434303 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-utilities\") pod \"community-operators-kr8hl\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.434851 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-utilities\") pod \"community-operators-kr8hl\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.435213 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-catalog-content\") pod \"community-operators-kr8hl\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.458462 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zxjg\" (UniqueName: \"kubernetes.io/projected/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-kube-api-access-5zxjg\") pod \"community-operators-kr8hl\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:07 crc kubenswrapper[4810]: I0930 09:12:07.608655 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:08 crc kubenswrapper[4810]: I0930 09:12:08.136955 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kr8hl"] Sep 30 09:12:08 crc kubenswrapper[4810]: I0930 09:12:08.985352 4810 generic.go:334] "Generic (PLEG): container finished" podID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" containerID="198dd8c914505f37f78784d11fd8adaa0c69057a060bd17c97a1699e81dd52bd" exitCode=0 Sep 30 09:12:08 crc kubenswrapper[4810]: I0930 09:12:08.985445 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr8hl" event={"ID":"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4","Type":"ContainerDied","Data":"198dd8c914505f37f78784d11fd8adaa0c69057a060bd17c97a1699e81dd52bd"} Sep 30 09:12:08 crc kubenswrapper[4810]: I0930 09:12:08.986080 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr8hl" event={"ID":"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4","Type":"ContainerStarted","Data":"5e000a5a339ef61c0dd23ead26c6ac8c391b96f32c64389033053b38d8c520be"} Sep 30 09:12:11 crc kubenswrapper[4810]: I0930 09:12:11.014496 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr8hl" event={"ID":"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4","Type":"ContainerStarted","Data":"3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970"} Sep 30 09:12:13 crc kubenswrapper[4810]: I0930 09:12:13.035322 4810 generic.go:334] "Generic (PLEG): container finished" podID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" containerID="3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970" exitCode=0 Sep 30 09:12:13 crc kubenswrapper[4810]: I0930 09:12:13.035387 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr8hl" event={"ID":"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4","Type":"ContainerDied","Data":"3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970"} Sep 30 09:12:13 crc kubenswrapper[4810]: I0930 09:12:13.307046 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:12:13 crc kubenswrapper[4810]: E0930 09:12:13.307553 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:12:14 crc kubenswrapper[4810]: I0930 09:12:14.050779 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr8hl" event={"ID":"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4","Type":"ContainerStarted","Data":"9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f"} Sep 30 09:12:14 crc kubenswrapper[4810]: I0930 09:12:14.072197 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kr8hl" podStartSLOduration=2.561910987 podStartE2EDuration="7.072174074s" podCreationTimestamp="2025-09-30 09:12:07 +0000 UTC" firstStartedPulling="2025-09-30 09:12:08.988230816 +0000 UTC m=+4152.440430093" lastFinishedPulling="2025-09-30 09:12:13.498493873 +0000 UTC m=+4156.950693180" observedRunningTime="2025-09-30 09:12:14.068452966 +0000 UTC m=+4157.520652233" watchObservedRunningTime="2025-09-30 09:12:14.072174074 +0000 UTC m=+4157.524373341" Sep 30 09:12:17 crc kubenswrapper[4810]: I0930 09:12:17.609691 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:17 crc kubenswrapper[4810]: I0930 09:12:17.610468 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:17 crc kubenswrapper[4810]: I0930 09:12:17.672799 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:18 crc kubenswrapper[4810]: I0930 09:12:18.181401 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:18 crc kubenswrapper[4810]: I0930 09:12:18.233197 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kr8hl"] Sep 30 09:12:20 crc kubenswrapper[4810]: I0930 09:12:20.120016 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kr8hl" podUID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" containerName="registry-server" containerID="cri-o://9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f" gracePeriod=2 Sep 30 09:12:20 crc kubenswrapper[4810]: I0930 09:12:20.684973 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:20 crc kubenswrapper[4810]: I0930 09:12:20.838114 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-catalog-content\") pod \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " Sep 30 09:12:20 crc kubenswrapper[4810]: I0930 09:12:20.838211 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zxjg\" (UniqueName: \"kubernetes.io/projected/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-kube-api-access-5zxjg\") pod \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " Sep 30 09:12:20 crc kubenswrapper[4810]: I0930 09:12:20.838520 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-utilities\") pod \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\" (UID: \"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4\") " Sep 30 09:12:20 crc kubenswrapper[4810]: I0930 09:12:20.839285 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-utilities" (OuterVolumeSpecName: "utilities") pod "c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" (UID: "c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:12:20 crc kubenswrapper[4810]: I0930 09:12:20.846980 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-kube-api-access-5zxjg" (OuterVolumeSpecName: "kube-api-access-5zxjg") pod "c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" (UID: "c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4"). InnerVolumeSpecName "kube-api-access-5zxjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:12:20 crc kubenswrapper[4810]: I0930 09:12:20.885770 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" (UID: "c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:12:20 crc kubenswrapper[4810]: I0930 09:12:20.941236 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:12:20 crc kubenswrapper[4810]: I0930 09:12:20.941307 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zxjg\" (UniqueName: \"kubernetes.io/projected/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-kube-api-access-5zxjg\") on node \"crc\" DevicePath \"\"" Sep 30 09:12:20 crc kubenswrapper[4810]: I0930 09:12:20.941327 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.161533 4810 generic.go:334] "Generic (PLEG): container finished" podID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" containerID="9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f" exitCode=0 Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.161589 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr8hl" event={"ID":"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4","Type":"ContainerDied","Data":"9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f"} Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.161619 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr8hl" event={"ID":"c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4","Type":"ContainerDied","Data":"5e000a5a339ef61c0dd23ead26c6ac8c391b96f32c64389033053b38d8c520be"} Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.161640 4810 scope.go:117] "RemoveContainer" containerID="9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f" Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.161811 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kr8hl" Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.190861 4810 scope.go:117] "RemoveContainer" containerID="3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970" Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.219398 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kr8hl"] Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.232511 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kr8hl"] Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.244361 4810 scope.go:117] "RemoveContainer" containerID="198dd8c914505f37f78784d11fd8adaa0c69057a060bd17c97a1699e81dd52bd" Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.283472 4810 scope.go:117] "RemoveContainer" containerID="9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f" Sep 30 09:12:21 crc kubenswrapper[4810]: E0930 09:12:21.285582 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f\": container with ID starting with 9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f not found: ID does not exist" containerID="9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f" Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.285621 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f"} err="failed to get container status \"9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f\": rpc error: code = NotFound desc = could not find container \"9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f\": container with ID starting with 9e4c8420eeddc9a7cd791c33cb9f32774723a96d586f5bf6c023621463d4b79f not found: ID does not exist" Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.285643 4810 scope.go:117] "RemoveContainer" containerID="3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970" Sep 30 09:12:21 crc kubenswrapper[4810]: E0930 09:12:21.287614 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970\": container with ID starting with 3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970 not found: ID does not exist" containerID="3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970" Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.287657 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970"} err="failed to get container status \"3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970\": rpc error: code = NotFound desc = could not find container \"3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970\": container with ID starting with 3f804cfd905c36403e73d7f128d32cb657bc246395f57e0ae304db9a17faf970 not found: ID does not exist" Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.287678 4810 scope.go:117] "RemoveContainer" containerID="198dd8c914505f37f78784d11fd8adaa0c69057a060bd17c97a1699e81dd52bd" Sep 30 09:12:21 crc kubenswrapper[4810]: E0930 09:12:21.288551 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"198dd8c914505f37f78784d11fd8adaa0c69057a060bd17c97a1699e81dd52bd\": container with ID starting with 198dd8c914505f37f78784d11fd8adaa0c69057a060bd17c97a1699e81dd52bd not found: ID does not exist" containerID="198dd8c914505f37f78784d11fd8adaa0c69057a060bd17c97a1699e81dd52bd" Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.288594 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"198dd8c914505f37f78784d11fd8adaa0c69057a060bd17c97a1699e81dd52bd"} err="failed to get container status \"198dd8c914505f37f78784d11fd8adaa0c69057a060bd17c97a1699e81dd52bd\": rpc error: code = NotFound desc = could not find container \"198dd8c914505f37f78784d11fd8adaa0c69057a060bd17c97a1699e81dd52bd\": container with ID starting with 198dd8c914505f37f78784d11fd8adaa0c69057a060bd17c97a1699e81dd52bd not found: ID does not exist" Sep 30 09:12:21 crc kubenswrapper[4810]: I0930 09:12:21.319536 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" path="/var/lib/kubelet/pods/c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4/volumes" Sep 30 09:12:24 crc kubenswrapper[4810]: I0930 09:12:24.307254 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:12:25 crc kubenswrapper[4810]: I0930 09:12:25.209960 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"48efc1b803f8a39be2b99333f2f5a344ef527a0f241cb6683a9b511799e44acf"} Sep 30 09:12:26 crc kubenswrapper[4810]: E0930 09:12:26.301637 4810 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.203:56590->38.102.83.203:42869: write tcp 38.102.83.203:56590->38.102.83.203:42869: write: broken pipe Sep 30 09:12:33 crc kubenswrapper[4810]: E0930 09:12:33.795374 4810 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.203:55844->38.102.83.203:42869: write tcp 38.102.83.203:55844->38.102.83.203:42869: write: connection reset by peer Sep 30 09:14:45 crc kubenswrapper[4810]: I0930 09:14:45.912326 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:14:45 crc kubenswrapper[4810]: I0930 09:14:45.912867 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.153814 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d"] Sep 30 09:15:00 crc kubenswrapper[4810]: E0930 09:15:00.156163 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" containerName="extract-content" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.156274 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" containerName="extract-content" Sep 30 09:15:00 crc kubenswrapper[4810]: E0930 09:15:00.156369 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" containerName="registry-server" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.156434 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" containerName="registry-server" Sep 30 09:15:00 crc kubenswrapper[4810]: E0930 09:15:00.156510 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" containerName="extract-utilities" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.156564 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" containerName="extract-utilities" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.157115 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3b0b36b-6ca0-476d-9988-2ecef5aa5ee4" containerName="registry-server" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.160693 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.167978 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.168684 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.171528 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d"] Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.241659 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3a23c9b-3307-48e3-8994-59c0271bbd44-config-volume\") pod \"collect-profiles-29320395-pfs9d\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.241743 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3a23c9b-3307-48e3-8994-59c0271bbd44-secret-volume\") pod \"collect-profiles-29320395-pfs9d\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.241864 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cb85\" (UniqueName: \"kubernetes.io/projected/a3a23c9b-3307-48e3-8994-59c0271bbd44-kube-api-access-4cb85\") pod \"collect-profiles-29320395-pfs9d\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.343851 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3a23c9b-3307-48e3-8994-59c0271bbd44-secret-volume\") pod \"collect-profiles-29320395-pfs9d\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.343961 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cb85\" (UniqueName: \"kubernetes.io/projected/a3a23c9b-3307-48e3-8994-59c0271bbd44-kube-api-access-4cb85\") pod \"collect-profiles-29320395-pfs9d\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.344209 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3a23c9b-3307-48e3-8994-59c0271bbd44-config-volume\") pod \"collect-profiles-29320395-pfs9d\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.344953 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3a23c9b-3307-48e3-8994-59c0271bbd44-config-volume\") pod \"collect-profiles-29320395-pfs9d\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.350178 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3a23c9b-3307-48e3-8994-59c0271bbd44-secret-volume\") pod \"collect-profiles-29320395-pfs9d\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.364348 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cb85\" (UniqueName: \"kubernetes.io/projected/a3a23c9b-3307-48e3-8994-59c0271bbd44-kube-api-access-4cb85\") pod \"collect-profiles-29320395-pfs9d\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.490902 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:00 crc kubenswrapper[4810]: I0930 09:15:00.951316 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d"] Sep 30 09:15:01 crc kubenswrapper[4810]: I0930 09:15:01.911760 4810 generic.go:334] "Generic (PLEG): container finished" podID="a3a23c9b-3307-48e3-8994-59c0271bbd44" containerID="821048672c7ea35284959168752ea6ff822a9f359e23dcc40ce10513929a05bb" exitCode=0 Sep 30 09:15:01 crc kubenswrapper[4810]: I0930 09:15:01.911840 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" event={"ID":"a3a23c9b-3307-48e3-8994-59c0271bbd44","Type":"ContainerDied","Data":"821048672c7ea35284959168752ea6ff822a9f359e23dcc40ce10513929a05bb"} Sep 30 09:15:01 crc kubenswrapper[4810]: I0930 09:15:01.912033 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" event={"ID":"a3a23c9b-3307-48e3-8994-59c0271bbd44","Type":"ContainerStarted","Data":"effc330b8f39b69fba8c570b30c22c34fb8a7c62385518cafb7110aa4c553bb5"} Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.283878 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.306046 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3a23c9b-3307-48e3-8994-59c0271bbd44-config-volume\") pod \"a3a23c9b-3307-48e3-8994-59c0271bbd44\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.306129 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cb85\" (UniqueName: \"kubernetes.io/projected/a3a23c9b-3307-48e3-8994-59c0271bbd44-kube-api-access-4cb85\") pod \"a3a23c9b-3307-48e3-8994-59c0271bbd44\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.306499 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3a23c9b-3307-48e3-8994-59c0271bbd44-secret-volume\") pod \"a3a23c9b-3307-48e3-8994-59c0271bbd44\" (UID: \"a3a23c9b-3307-48e3-8994-59c0271bbd44\") " Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.307313 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3a23c9b-3307-48e3-8994-59c0271bbd44-config-volume" (OuterVolumeSpecName: "config-volume") pod "a3a23c9b-3307-48e3-8994-59c0271bbd44" (UID: "a3a23c9b-3307-48e3-8994-59c0271bbd44"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.313029 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3a23c9b-3307-48e3-8994-59c0271bbd44-kube-api-access-4cb85" (OuterVolumeSpecName: "kube-api-access-4cb85") pod "a3a23c9b-3307-48e3-8994-59c0271bbd44" (UID: "a3a23c9b-3307-48e3-8994-59c0271bbd44"). InnerVolumeSpecName "kube-api-access-4cb85". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.313472 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3a23c9b-3307-48e3-8994-59c0271bbd44-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a3a23c9b-3307-48e3-8994-59c0271bbd44" (UID: "a3a23c9b-3307-48e3-8994-59c0271bbd44"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.408747 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3a23c9b-3307-48e3-8994-59c0271bbd44-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.408786 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3a23c9b-3307-48e3-8994-59c0271bbd44-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.408800 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cb85\" (UniqueName: \"kubernetes.io/projected/a3a23c9b-3307-48e3-8994-59c0271bbd44-kube-api-access-4cb85\") on node \"crc\" DevicePath \"\"" Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.931474 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" event={"ID":"a3a23c9b-3307-48e3-8994-59c0271bbd44","Type":"ContainerDied","Data":"effc330b8f39b69fba8c570b30c22c34fb8a7c62385518cafb7110aa4c553bb5"} Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.931995 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="effc330b8f39b69fba8c570b30c22c34fb8a7c62385518cafb7110aa4c553bb5" Sep 30 09:15:03 crc kubenswrapper[4810]: I0930 09:15:03.931551 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320395-pfs9d" Sep 30 09:15:04 crc kubenswrapper[4810]: I0930 09:15:04.361125 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2"] Sep 30 09:15:04 crc kubenswrapper[4810]: I0930 09:15:04.378618 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320350-jptm2"] Sep 30 09:15:05 crc kubenswrapper[4810]: I0930 09:15:05.323652 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3619e2ef-5db9-4f4a-b8f3-95f93fdba325" path="/var/lib/kubelet/pods/3619e2ef-5db9-4f4a-b8f3-95f93fdba325/volumes" Sep 30 09:15:15 crc kubenswrapper[4810]: I0930 09:15:15.912642 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:15:15 crc kubenswrapper[4810]: I0930 09:15:15.913212 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:15:22 crc kubenswrapper[4810]: I0930 09:15:22.613711 4810 scope.go:117] "RemoveContainer" containerID="2f23a6139b5411612599b0bd17ef1b314babb37e4d6c38927f7adbec1a400e57" Sep 30 09:15:22 crc kubenswrapper[4810]: I0930 09:15:22.664469 4810 scope.go:117] "RemoveContainer" containerID="e507a662df5b054b8fd2245360b2f0928a7d1334ea97d86425fd12a60a14b512" Sep 30 09:15:22 crc kubenswrapper[4810]: I0930 09:15:22.697458 4810 scope.go:117] "RemoveContainer" containerID="d5a115b07989ccc126b6d6692af6d30c6177610fe11065984fa52c8d55f42476" Sep 30 09:15:22 crc kubenswrapper[4810]: I0930 09:15:22.736756 4810 scope.go:117] "RemoveContainer" containerID="72f3ecb1f173b409d561d4600033cc578422f5da5ae47261d8bc28a6c407cad9" Sep 30 09:15:45 crc kubenswrapper[4810]: I0930 09:15:45.912328 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:15:45 crc kubenswrapper[4810]: I0930 09:15:45.912883 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:15:45 crc kubenswrapper[4810]: I0930 09:15:45.912945 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 09:15:45 crc kubenswrapper[4810]: I0930 09:15:45.913812 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"48efc1b803f8a39be2b99333f2f5a344ef527a0f241cb6683a9b511799e44acf"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 09:15:45 crc kubenswrapper[4810]: I0930 09:15:45.913871 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://48efc1b803f8a39be2b99333f2f5a344ef527a0f241cb6683a9b511799e44acf" gracePeriod=600 Sep 30 09:15:46 crc kubenswrapper[4810]: I0930 09:15:46.363604 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="48efc1b803f8a39be2b99333f2f5a344ef527a0f241cb6683a9b511799e44acf" exitCode=0 Sep 30 09:15:46 crc kubenswrapper[4810]: I0930 09:15:46.363691 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"48efc1b803f8a39be2b99333f2f5a344ef527a0f241cb6683a9b511799e44acf"} Sep 30 09:15:46 crc kubenswrapper[4810]: I0930 09:15:46.364213 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f"} Sep 30 09:15:46 crc kubenswrapper[4810]: I0930 09:15:46.364273 4810 scope.go:117] "RemoveContainer" containerID="241106060b2863c809e7ac6395be6079ef9d6b93fbf6cca102aae467b961a83f" Sep 30 09:17:47 crc kubenswrapper[4810]: I0930 09:17:47.784890 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g9qn5"] Sep 30 09:17:47 crc kubenswrapper[4810]: E0930 09:17:47.789766 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a23c9b-3307-48e3-8994-59c0271bbd44" containerName="collect-profiles" Sep 30 09:17:47 crc kubenswrapper[4810]: I0930 09:17:47.789811 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a23c9b-3307-48e3-8994-59c0271bbd44" containerName="collect-profiles" Sep 30 09:17:47 crc kubenswrapper[4810]: I0930 09:17:47.790199 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a23c9b-3307-48e3-8994-59c0271bbd44" containerName="collect-profiles" Sep 30 09:17:47 crc kubenswrapper[4810]: I0930 09:17:47.792155 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:47 crc kubenswrapper[4810]: I0930 09:17:47.803606 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g9qn5"] Sep 30 09:17:47 crc kubenswrapper[4810]: I0930 09:17:47.970702 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-utilities\") pod \"redhat-marketplace-g9qn5\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:47 crc kubenswrapper[4810]: I0930 09:17:47.970761 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlqn2\" (UniqueName: \"kubernetes.io/projected/5fd50b79-2aae-4ff2-974a-c693b9969c73-kube-api-access-nlqn2\") pod \"redhat-marketplace-g9qn5\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:47 crc kubenswrapper[4810]: I0930 09:17:47.970787 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-catalog-content\") pod \"redhat-marketplace-g9qn5\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:48 crc kubenswrapper[4810]: I0930 09:17:48.073682 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-utilities\") pod \"redhat-marketplace-g9qn5\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:48 crc kubenswrapper[4810]: I0930 09:17:48.073727 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlqn2\" (UniqueName: \"kubernetes.io/projected/5fd50b79-2aae-4ff2-974a-c693b9969c73-kube-api-access-nlqn2\") pod \"redhat-marketplace-g9qn5\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:48 crc kubenswrapper[4810]: I0930 09:17:48.073755 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-catalog-content\") pod \"redhat-marketplace-g9qn5\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:48 crc kubenswrapper[4810]: I0930 09:17:48.074285 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-utilities\") pod \"redhat-marketplace-g9qn5\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:48 crc kubenswrapper[4810]: I0930 09:17:48.074567 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-catalog-content\") pod \"redhat-marketplace-g9qn5\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:48 crc kubenswrapper[4810]: I0930 09:17:48.100332 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlqn2\" (UniqueName: \"kubernetes.io/projected/5fd50b79-2aae-4ff2-974a-c693b9969c73-kube-api-access-nlqn2\") pod \"redhat-marketplace-g9qn5\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:48 crc kubenswrapper[4810]: I0930 09:17:48.128580 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:48 crc kubenswrapper[4810]: I0930 09:17:48.648444 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g9qn5"] Sep 30 09:17:48 crc kubenswrapper[4810]: I0930 09:17:48.688297 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9qn5" event={"ID":"5fd50b79-2aae-4ff2-974a-c693b9969c73","Type":"ContainerStarted","Data":"cbadfaba69cca02dd6c8e6d4a84c045dfd888afc5ca9c8f35bef3ce7764d3415"} Sep 30 09:17:49 crc kubenswrapper[4810]: I0930 09:17:49.703681 4810 generic.go:334] "Generic (PLEG): container finished" podID="5fd50b79-2aae-4ff2-974a-c693b9969c73" containerID="71fd716175841abe80939ac40ae534a341823b8244d0e6b0431db4bbfff1ddcb" exitCode=0 Sep 30 09:17:49 crc kubenswrapper[4810]: I0930 09:17:49.703759 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9qn5" event={"ID":"5fd50b79-2aae-4ff2-974a-c693b9969c73","Type":"ContainerDied","Data":"71fd716175841abe80939ac40ae534a341823b8244d0e6b0431db4bbfff1ddcb"} Sep 30 09:17:49 crc kubenswrapper[4810]: I0930 09:17:49.706647 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 09:17:50 crc kubenswrapper[4810]: I0930 09:17:50.719219 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9qn5" event={"ID":"5fd50b79-2aae-4ff2-974a-c693b9969c73","Type":"ContainerStarted","Data":"c9defc48a96e21fd32e5547dfeb1b80a59a6fd3a1fc9613e9731cdfd16d7b110"} Sep 30 09:17:51 crc kubenswrapper[4810]: I0930 09:17:51.735164 4810 generic.go:334] "Generic (PLEG): container finished" podID="5fd50b79-2aae-4ff2-974a-c693b9969c73" containerID="c9defc48a96e21fd32e5547dfeb1b80a59a6fd3a1fc9613e9731cdfd16d7b110" exitCode=0 Sep 30 09:17:51 crc kubenswrapper[4810]: I0930 09:17:51.735237 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9qn5" event={"ID":"5fd50b79-2aae-4ff2-974a-c693b9969c73","Type":"ContainerDied","Data":"c9defc48a96e21fd32e5547dfeb1b80a59a6fd3a1fc9613e9731cdfd16d7b110"} Sep 30 09:17:51 crc kubenswrapper[4810]: I0930 09:17:51.735537 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9qn5" event={"ID":"5fd50b79-2aae-4ff2-974a-c693b9969c73","Type":"ContainerStarted","Data":"88d49d162ee4e7858948d733775f0f227c9402b5e2cad421371fc35a21c5b073"} Sep 30 09:17:51 crc kubenswrapper[4810]: I0930 09:17:51.764225 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g9qn5" podStartSLOduration=3.30791869 podStartE2EDuration="4.764205106s" podCreationTimestamp="2025-09-30 09:17:47 +0000 UTC" firstStartedPulling="2025-09-30 09:17:49.706420265 +0000 UTC m=+4493.158619532" lastFinishedPulling="2025-09-30 09:17:51.162706671 +0000 UTC m=+4494.614905948" observedRunningTime="2025-09-30 09:17:51.758726197 +0000 UTC m=+4495.210925464" watchObservedRunningTime="2025-09-30 09:17:51.764205106 +0000 UTC m=+4495.216404373" Sep 30 09:17:58 crc kubenswrapper[4810]: I0930 09:17:58.129510 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:58 crc kubenswrapper[4810]: I0930 09:17:58.131109 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:58 crc kubenswrapper[4810]: I0930 09:17:58.242019 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:58 crc kubenswrapper[4810]: I0930 09:17:58.915561 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:17:58 crc kubenswrapper[4810]: I0930 09:17:58.978374 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g9qn5"] Sep 30 09:18:00 crc kubenswrapper[4810]: I0930 09:18:00.847811 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g9qn5" podUID="5fd50b79-2aae-4ff2-974a-c693b9969c73" containerName="registry-server" containerID="cri-o://88d49d162ee4e7858948d733775f0f227c9402b5e2cad421371fc35a21c5b073" gracePeriod=2 Sep 30 09:18:01 crc kubenswrapper[4810]: I0930 09:18:01.862643 4810 generic.go:334] "Generic (PLEG): container finished" podID="5fd50b79-2aae-4ff2-974a-c693b9969c73" containerID="88d49d162ee4e7858948d733775f0f227c9402b5e2cad421371fc35a21c5b073" exitCode=0 Sep 30 09:18:01 crc kubenswrapper[4810]: I0930 09:18:01.862758 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9qn5" event={"ID":"5fd50b79-2aae-4ff2-974a-c693b9969c73","Type":"ContainerDied","Data":"88d49d162ee4e7858948d733775f0f227c9402b5e2cad421371fc35a21c5b073"} Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.082595 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.172942 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-utilities\") pod \"5fd50b79-2aae-4ff2-974a-c693b9969c73\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.173804 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-utilities" (OuterVolumeSpecName: "utilities") pod "5fd50b79-2aae-4ff2-974a-c693b9969c73" (UID: "5fd50b79-2aae-4ff2-974a-c693b9969c73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.173855 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-catalog-content\") pod \"5fd50b79-2aae-4ff2-974a-c693b9969c73\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.174012 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlqn2\" (UniqueName: \"kubernetes.io/projected/5fd50b79-2aae-4ff2-974a-c693b9969c73-kube-api-access-nlqn2\") pod \"5fd50b79-2aae-4ff2-974a-c693b9969c73\" (UID: \"5fd50b79-2aae-4ff2-974a-c693b9969c73\") " Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.174961 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.182627 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd50b79-2aae-4ff2-974a-c693b9969c73-kube-api-access-nlqn2" (OuterVolumeSpecName: "kube-api-access-nlqn2") pod "5fd50b79-2aae-4ff2-974a-c693b9969c73" (UID: "5fd50b79-2aae-4ff2-974a-c693b9969c73"). InnerVolumeSpecName "kube-api-access-nlqn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.185766 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fd50b79-2aae-4ff2-974a-c693b9969c73" (UID: "5fd50b79-2aae-4ff2-974a-c693b9969c73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.277036 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlqn2\" (UniqueName: \"kubernetes.io/projected/5fd50b79-2aae-4ff2-974a-c693b9969c73-kube-api-access-nlqn2\") on node \"crc\" DevicePath \"\"" Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.277082 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fd50b79-2aae-4ff2-974a-c693b9969c73-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.879230 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9qn5" event={"ID":"5fd50b79-2aae-4ff2-974a-c693b9969c73","Type":"ContainerDied","Data":"cbadfaba69cca02dd6c8e6d4a84c045dfd888afc5ca9c8f35bef3ce7764d3415"} Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.879378 4810 scope.go:117] "RemoveContainer" containerID="88d49d162ee4e7858948d733775f0f227c9402b5e2cad421371fc35a21c5b073" Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.879399 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g9qn5" Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.905090 4810 scope.go:117] "RemoveContainer" containerID="c9defc48a96e21fd32e5547dfeb1b80a59a6fd3a1fc9613e9731cdfd16d7b110" Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.929016 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g9qn5"] Sep 30 09:18:02 crc kubenswrapper[4810]: I0930 09:18:02.943101 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g9qn5"] Sep 30 09:18:03 crc kubenswrapper[4810]: I0930 09:18:03.326666 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fd50b79-2aae-4ff2-974a-c693b9969c73" path="/var/lib/kubelet/pods/5fd50b79-2aae-4ff2-974a-c693b9969c73/volumes" Sep 30 09:18:03 crc kubenswrapper[4810]: I0930 09:18:03.417189 4810 scope.go:117] "RemoveContainer" containerID="71fd716175841abe80939ac40ae534a341823b8244d0e6b0431db4bbfff1ddcb" Sep 30 09:18:15 crc kubenswrapper[4810]: I0930 09:18:15.911173 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:18:15 crc kubenswrapper[4810]: I0930 09:18:15.911844 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:18:45 crc kubenswrapper[4810]: I0930 09:18:45.911597 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:18:45 crc kubenswrapper[4810]: I0930 09:18:45.912297 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:19:15 crc kubenswrapper[4810]: I0930 09:19:15.911854 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:19:15 crc kubenswrapper[4810]: I0930 09:19:15.912562 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:19:15 crc kubenswrapper[4810]: I0930 09:19:15.912612 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 09:19:15 crc kubenswrapper[4810]: I0930 09:19:15.913476 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 09:19:15 crc kubenswrapper[4810]: I0930 09:19:15.913536 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" gracePeriod=600 Sep 30 09:19:16 crc kubenswrapper[4810]: E0930 09:19:16.639531 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:19:16 crc kubenswrapper[4810]: I0930 09:19:16.691134 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" exitCode=0 Sep 30 09:19:16 crc kubenswrapper[4810]: I0930 09:19:16.691185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f"} Sep 30 09:19:16 crc kubenswrapper[4810]: I0930 09:19:16.691223 4810 scope.go:117] "RemoveContainer" containerID="48efc1b803f8a39be2b99333f2f5a344ef527a0f241cb6683a9b511799e44acf" Sep 30 09:19:16 crc kubenswrapper[4810]: I0930 09:19:16.691997 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:19:16 crc kubenswrapper[4810]: E0930 09:19:16.692360 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:19:30 crc kubenswrapper[4810]: I0930 09:19:30.306953 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:19:30 crc kubenswrapper[4810]: E0930 09:19:30.307850 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.419775 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j6xgs"] Sep 30 09:19:31 crc kubenswrapper[4810]: E0930 09:19:31.420642 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd50b79-2aae-4ff2-974a-c693b9969c73" containerName="extract-content" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.420658 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd50b79-2aae-4ff2-974a-c693b9969c73" containerName="extract-content" Sep 30 09:19:31 crc kubenswrapper[4810]: E0930 09:19:31.420683 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd50b79-2aae-4ff2-974a-c693b9969c73" containerName="extract-utilities" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.420691 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd50b79-2aae-4ff2-974a-c693b9969c73" containerName="extract-utilities" Sep 30 09:19:31 crc kubenswrapper[4810]: E0930 09:19:31.420714 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd50b79-2aae-4ff2-974a-c693b9969c73" containerName="registry-server" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.420722 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd50b79-2aae-4ff2-974a-c693b9969c73" containerName="registry-server" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.420961 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd50b79-2aae-4ff2-974a-c693b9969c73" containerName="registry-server" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.422391 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.442115 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j6xgs"] Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.588995 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-catalog-content\") pod \"certified-operators-j6xgs\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.589113 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz9hx\" (UniqueName: \"kubernetes.io/projected/300b645a-3bec-4061-8c88-4d5cd0f2496f-kube-api-access-lz9hx\") pod \"certified-operators-j6xgs\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.589149 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-utilities\") pod \"certified-operators-j6xgs\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.690834 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-catalog-content\") pod \"certified-operators-j6xgs\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.691236 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz9hx\" (UniqueName: \"kubernetes.io/projected/300b645a-3bec-4061-8c88-4d5cd0f2496f-kube-api-access-lz9hx\") pod \"certified-operators-j6xgs\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.691387 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-utilities\") pod \"certified-operators-j6xgs\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.691309 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-catalog-content\") pod \"certified-operators-j6xgs\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.692193 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-utilities\") pod \"certified-operators-j6xgs\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:31 crc kubenswrapper[4810]: I0930 09:19:31.766391 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz9hx\" (UniqueName: \"kubernetes.io/projected/300b645a-3bec-4061-8c88-4d5cd0f2496f-kube-api-access-lz9hx\") pod \"certified-operators-j6xgs\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:32 crc kubenswrapper[4810]: I0930 09:19:32.051707 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:32 crc kubenswrapper[4810]: I0930 09:19:32.601697 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j6xgs"] Sep 30 09:19:32 crc kubenswrapper[4810]: I0930 09:19:32.887985 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6xgs" event={"ID":"300b645a-3bec-4061-8c88-4d5cd0f2496f","Type":"ContainerStarted","Data":"2ffbfe1a7e595e5ef0273d4f9ce5b5756acc92d11caec7b7a8bd32776eb5225b"} Sep 30 09:19:32 crc kubenswrapper[4810]: I0930 09:19:32.888577 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6xgs" event={"ID":"300b645a-3bec-4061-8c88-4d5cd0f2496f","Type":"ContainerStarted","Data":"63b01f31c45c56c9ab9fdc894968aaac322c800d3c00bcc176ded2650b0698b1"} Sep 30 09:19:33 crc kubenswrapper[4810]: I0930 09:19:33.915594 4810 generic.go:334] "Generic (PLEG): container finished" podID="300b645a-3bec-4061-8c88-4d5cd0f2496f" containerID="2ffbfe1a7e595e5ef0273d4f9ce5b5756acc92d11caec7b7a8bd32776eb5225b" exitCode=0 Sep 30 09:19:33 crc kubenswrapper[4810]: I0930 09:19:33.915672 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6xgs" event={"ID":"300b645a-3bec-4061-8c88-4d5cd0f2496f","Type":"ContainerDied","Data":"2ffbfe1a7e595e5ef0273d4f9ce5b5756acc92d11caec7b7a8bd32776eb5225b"} Sep 30 09:19:35 crc kubenswrapper[4810]: I0930 09:19:35.940596 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6xgs" event={"ID":"300b645a-3bec-4061-8c88-4d5cd0f2496f","Type":"ContainerStarted","Data":"58fe6c56f82efaa74c83814e79efdc97044e8557821e47bf5508bff6543c7d3d"} Sep 30 09:19:36 crc kubenswrapper[4810]: I0930 09:19:36.829546 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mxj9j"] Sep 30 09:19:36 crc kubenswrapper[4810]: I0930 09:19:36.832206 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:36 crc kubenswrapper[4810]: I0930 09:19:36.840705 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mxj9j"] Sep 30 09:19:36 crc kubenswrapper[4810]: I0930 09:19:36.915395 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsrnt\" (UniqueName: \"kubernetes.io/projected/31c34400-0ef5-4999-b235-582857bc9f5d-kube-api-access-rsrnt\") pod \"redhat-operators-mxj9j\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:36 crc kubenswrapper[4810]: I0930 09:19:36.915590 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-utilities\") pod \"redhat-operators-mxj9j\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:36 crc kubenswrapper[4810]: I0930 09:19:36.915849 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-catalog-content\") pod \"redhat-operators-mxj9j\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:36 crc kubenswrapper[4810]: I0930 09:19:36.952881 4810 generic.go:334] "Generic (PLEG): container finished" podID="300b645a-3bec-4061-8c88-4d5cd0f2496f" containerID="58fe6c56f82efaa74c83814e79efdc97044e8557821e47bf5508bff6543c7d3d" exitCode=0 Sep 30 09:19:36 crc kubenswrapper[4810]: I0930 09:19:36.953085 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6xgs" event={"ID":"300b645a-3bec-4061-8c88-4d5cd0f2496f","Type":"ContainerDied","Data":"58fe6c56f82efaa74c83814e79efdc97044e8557821e47bf5508bff6543c7d3d"} Sep 30 09:19:37 crc kubenswrapper[4810]: I0930 09:19:37.017950 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsrnt\" (UniqueName: \"kubernetes.io/projected/31c34400-0ef5-4999-b235-582857bc9f5d-kube-api-access-rsrnt\") pod \"redhat-operators-mxj9j\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:37 crc kubenswrapper[4810]: I0930 09:19:37.018094 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-utilities\") pod \"redhat-operators-mxj9j\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:37 crc kubenswrapper[4810]: I0930 09:19:37.018230 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-catalog-content\") pod \"redhat-operators-mxj9j\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:37 crc kubenswrapper[4810]: I0930 09:19:37.018887 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-catalog-content\") pod \"redhat-operators-mxj9j\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:37 crc kubenswrapper[4810]: I0930 09:19:37.022138 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-utilities\") pod \"redhat-operators-mxj9j\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:37 crc kubenswrapper[4810]: I0930 09:19:37.045960 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsrnt\" (UniqueName: \"kubernetes.io/projected/31c34400-0ef5-4999-b235-582857bc9f5d-kube-api-access-rsrnt\") pod \"redhat-operators-mxj9j\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:37 crc kubenswrapper[4810]: I0930 09:19:37.161590 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:37 crc kubenswrapper[4810]: I0930 09:19:37.663405 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mxj9j"] Sep 30 09:19:37 crc kubenswrapper[4810]: W0930 09:19:37.667403 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31c34400_0ef5_4999_b235_582857bc9f5d.slice/crio-ef540095e88f0d48be0899bfffacf906019bf7e31daa7d4a9c65526c24757746 WatchSource:0}: Error finding container ef540095e88f0d48be0899bfffacf906019bf7e31daa7d4a9c65526c24757746: Status 404 returned error can't find the container with id ef540095e88f0d48be0899bfffacf906019bf7e31daa7d4a9c65526c24757746 Sep 30 09:19:37 crc kubenswrapper[4810]: I0930 09:19:37.962620 4810 generic.go:334] "Generic (PLEG): container finished" podID="31c34400-0ef5-4999-b235-582857bc9f5d" containerID="32ed01cbcf727397fa78858435db66d329e9c9c872236ef5190018c3b358cca7" exitCode=0 Sep 30 09:19:37 crc kubenswrapper[4810]: I0930 09:19:37.962659 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxj9j" event={"ID":"31c34400-0ef5-4999-b235-582857bc9f5d","Type":"ContainerDied","Data":"32ed01cbcf727397fa78858435db66d329e9c9c872236ef5190018c3b358cca7"} Sep 30 09:19:37 crc kubenswrapper[4810]: I0930 09:19:37.962683 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxj9j" event={"ID":"31c34400-0ef5-4999-b235-582857bc9f5d","Type":"ContainerStarted","Data":"ef540095e88f0d48be0899bfffacf906019bf7e31daa7d4a9c65526c24757746"} Sep 30 09:19:38 crc kubenswrapper[4810]: I0930 09:19:38.975442 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6xgs" event={"ID":"300b645a-3bec-4061-8c88-4d5cd0f2496f","Type":"ContainerStarted","Data":"10cab7ac197f5aa1cedbac3c632f36d1345623ac402d380ec75c4f43a2d19a37"} Sep 30 09:19:39 crc kubenswrapper[4810]: I0930 09:19:39.005678 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j6xgs" podStartSLOduration=4.083083061 podStartE2EDuration="8.005652134s" podCreationTimestamp="2025-09-30 09:19:31 +0000 UTC" firstStartedPulling="2025-09-30 09:19:33.920160115 +0000 UTC m=+4597.372359422" lastFinishedPulling="2025-09-30 09:19:37.842729228 +0000 UTC m=+4601.294928495" observedRunningTime="2025-09-30 09:19:39.004365096 +0000 UTC m=+4602.456564373" watchObservedRunningTime="2025-09-30 09:19:39.005652134 +0000 UTC m=+4602.457851421" Sep 30 09:19:39 crc kubenswrapper[4810]: I0930 09:19:39.988333 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxj9j" event={"ID":"31c34400-0ef5-4999-b235-582857bc9f5d","Type":"ContainerStarted","Data":"7f12475ee4f81c4467fecc2eab96e6b2b2c38b24a4d34aaf1368b789092c00eb"} Sep 30 09:19:41 crc kubenswrapper[4810]: I0930 09:19:41.008028 4810 generic.go:334] "Generic (PLEG): container finished" podID="31c34400-0ef5-4999-b235-582857bc9f5d" containerID="7f12475ee4f81c4467fecc2eab96e6b2b2c38b24a4d34aaf1368b789092c00eb" exitCode=0 Sep 30 09:19:41 crc kubenswrapper[4810]: I0930 09:19:41.008014 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxj9j" event={"ID":"31c34400-0ef5-4999-b235-582857bc9f5d","Type":"ContainerDied","Data":"7f12475ee4f81c4467fecc2eab96e6b2b2c38b24a4d34aaf1368b789092c00eb"} Sep 30 09:19:42 crc kubenswrapper[4810]: I0930 09:19:42.020770 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxj9j" event={"ID":"31c34400-0ef5-4999-b235-582857bc9f5d","Type":"ContainerStarted","Data":"fe4c9acc4cc06939f7c9800f708977d75dd88605c76a4f544000a56b174635bf"} Sep 30 09:19:42 crc kubenswrapper[4810]: I0930 09:19:42.050963 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mxj9j" podStartSLOduration=2.4034981269999998 podStartE2EDuration="6.050948028s" podCreationTimestamp="2025-09-30 09:19:36 +0000 UTC" firstStartedPulling="2025-09-30 09:19:37.964376519 +0000 UTC m=+4601.416575786" lastFinishedPulling="2025-09-30 09:19:41.61182641 +0000 UTC m=+4605.064025687" observedRunningTime="2025-09-30 09:19:42.045592213 +0000 UTC m=+4605.497791480" watchObservedRunningTime="2025-09-30 09:19:42.050948028 +0000 UTC m=+4605.503147295" Sep 30 09:19:42 crc kubenswrapper[4810]: I0930 09:19:42.052735 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:42 crc kubenswrapper[4810]: I0930 09:19:42.052796 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:42 crc kubenswrapper[4810]: I0930 09:19:42.103912 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:43 crc kubenswrapper[4810]: I0930 09:19:43.091782 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:43 crc kubenswrapper[4810]: I0930 09:19:43.307164 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:19:43 crc kubenswrapper[4810]: E0930 09:19:43.307650 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:19:45 crc kubenswrapper[4810]: I0930 09:19:45.797525 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j6xgs"] Sep 30 09:19:45 crc kubenswrapper[4810]: I0930 09:19:45.798032 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j6xgs" podUID="300b645a-3bec-4061-8c88-4d5cd0f2496f" containerName="registry-server" containerID="cri-o://10cab7ac197f5aa1cedbac3c632f36d1345623ac402d380ec75c4f43a2d19a37" gracePeriod=2 Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.091704 4810 generic.go:334] "Generic (PLEG): container finished" podID="300b645a-3bec-4061-8c88-4d5cd0f2496f" containerID="10cab7ac197f5aa1cedbac3c632f36d1345623ac402d380ec75c4f43a2d19a37" exitCode=0 Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.091796 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6xgs" event={"ID":"300b645a-3bec-4061-8c88-4d5cd0f2496f","Type":"ContainerDied","Data":"10cab7ac197f5aa1cedbac3c632f36d1345623ac402d380ec75c4f43a2d19a37"} Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.352372 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.413700 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9hx\" (UniqueName: \"kubernetes.io/projected/300b645a-3bec-4061-8c88-4d5cd0f2496f-kube-api-access-lz9hx\") pod \"300b645a-3bec-4061-8c88-4d5cd0f2496f\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.413934 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-catalog-content\") pod \"300b645a-3bec-4061-8c88-4d5cd0f2496f\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.413991 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-utilities\") pod \"300b645a-3bec-4061-8c88-4d5cd0f2496f\" (UID: \"300b645a-3bec-4061-8c88-4d5cd0f2496f\") " Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.415883 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-utilities" (OuterVolumeSpecName: "utilities") pod "300b645a-3bec-4061-8c88-4d5cd0f2496f" (UID: "300b645a-3bec-4061-8c88-4d5cd0f2496f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.420373 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/300b645a-3bec-4061-8c88-4d5cd0f2496f-kube-api-access-lz9hx" (OuterVolumeSpecName: "kube-api-access-lz9hx") pod "300b645a-3bec-4061-8c88-4d5cd0f2496f" (UID: "300b645a-3bec-4061-8c88-4d5cd0f2496f"). InnerVolumeSpecName "kube-api-access-lz9hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.461666 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "300b645a-3bec-4061-8c88-4d5cd0f2496f" (UID: "300b645a-3bec-4061-8c88-4d5cd0f2496f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.516566 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.516600 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/300b645a-3bec-4061-8c88-4d5cd0f2496f-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:19:46 crc kubenswrapper[4810]: I0930 09:19:46.516615 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9hx\" (UniqueName: \"kubernetes.io/projected/300b645a-3bec-4061-8c88-4d5cd0f2496f-kube-api-access-lz9hx\") on node \"crc\" DevicePath \"\"" Sep 30 09:19:47 crc kubenswrapper[4810]: I0930 09:19:47.104118 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6xgs" event={"ID":"300b645a-3bec-4061-8c88-4d5cd0f2496f","Type":"ContainerDied","Data":"63b01f31c45c56c9ab9fdc894968aaac322c800d3c00bcc176ded2650b0698b1"} Sep 30 09:19:47 crc kubenswrapper[4810]: I0930 09:19:47.104378 4810 scope.go:117] "RemoveContainer" containerID="10cab7ac197f5aa1cedbac3c632f36d1345623ac402d380ec75c4f43a2d19a37" Sep 30 09:19:47 crc kubenswrapper[4810]: I0930 09:19:47.104191 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6xgs" Sep 30 09:19:47 crc kubenswrapper[4810]: I0930 09:19:47.139715 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j6xgs"] Sep 30 09:19:47 crc kubenswrapper[4810]: I0930 09:19:47.147870 4810 scope.go:117] "RemoveContainer" containerID="58fe6c56f82efaa74c83814e79efdc97044e8557821e47bf5508bff6543c7d3d" Sep 30 09:19:47 crc kubenswrapper[4810]: I0930 09:19:47.154878 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j6xgs"] Sep 30 09:19:47 crc kubenswrapper[4810]: I0930 09:19:47.162512 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:47 crc kubenswrapper[4810]: I0930 09:19:47.162773 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:47 crc kubenswrapper[4810]: I0930 09:19:47.170889 4810 scope.go:117] "RemoveContainer" containerID="2ffbfe1a7e595e5ef0273d4f9ce5b5756acc92d11caec7b7a8bd32776eb5225b" Sep 30 09:19:47 crc kubenswrapper[4810]: I0930 09:19:47.233874 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:47 crc kubenswrapper[4810]: I0930 09:19:47.317028 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="300b645a-3bec-4061-8c88-4d5cd0f2496f" path="/var/lib/kubelet/pods/300b645a-3bec-4061-8c88-4d5cd0f2496f/volumes" Sep 30 09:19:48 crc kubenswrapper[4810]: I0930 09:19:48.176397 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:49 crc kubenswrapper[4810]: I0930 09:19:49.609866 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mxj9j"] Sep 30 09:19:51 crc kubenswrapper[4810]: I0930 09:19:51.145292 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mxj9j" podUID="31c34400-0ef5-4999-b235-582857bc9f5d" containerName="registry-server" containerID="cri-o://fe4c9acc4cc06939f7c9800f708977d75dd88605c76a4f544000a56b174635bf" gracePeriod=2 Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.166837 4810 generic.go:334] "Generic (PLEG): container finished" podID="31c34400-0ef5-4999-b235-582857bc9f5d" containerID="fe4c9acc4cc06939f7c9800f708977d75dd88605c76a4f544000a56b174635bf" exitCode=0 Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.166938 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxj9j" event={"ID":"31c34400-0ef5-4999-b235-582857bc9f5d","Type":"ContainerDied","Data":"fe4c9acc4cc06939f7c9800f708977d75dd88605c76a4f544000a56b174635bf"} Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.296649 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.333014 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-utilities\") pod \"31c34400-0ef5-4999-b235-582857bc9f5d\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.333097 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-catalog-content\") pod \"31c34400-0ef5-4999-b235-582857bc9f5d\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.333128 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsrnt\" (UniqueName: \"kubernetes.io/projected/31c34400-0ef5-4999-b235-582857bc9f5d-kube-api-access-rsrnt\") pod \"31c34400-0ef5-4999-b235-582857bc9f5d\" (UID: \"31c34400-0ef5-4999-b235-582857bc9f5d\") " Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.335144 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-utilities" (OuterVolumeSpecName: "utilities") pod "31c34400-0ef5-4999-b235-582857bc9f5d" (UID: "31c34400-0ef5-4999-b235-582857bc9f5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.343563 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31c34400-0ef5-4999-b235-582857bc9f5d-kube-api-access-rsrnt" (OuterVolumeSpecName: "kube-api-access-rsrnt") pod "31c34400-0ef5-4999-b235-582857bc9f5d" (UID: "31c34400-0ef5-4999-b235-582857bc9f5d"). InnerVolumeSpecName "kube-api-access-rsrnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.419988 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31c34400-0ef5-4999-b235-582857bc9f5d" (UID: "31c34400-0ef5-4999-b235-582857bc9f5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.435805 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.435848 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31c34400-0ef5-4999-b235-582857bc9f5d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:19:52 crc kubenswrapper[4810]: I0930 09:19:52.435859 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsrnt\" (UniqueName: \"kubernetes.io/projected/31c34400-0ef5-4999-b235-582857bc9f5d-kube-api-access-rsrnt\") on node \"crc\" DevicePath \"\"" Sep 30 09:19:53 crc kubenswrapper[4810]: I0930 09:19:53.181023 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxj9j" event={"ID":"31c34400-0ef5-4999-b235-582857bc9f5d","Type":"ContainerDied","Data":"ef540095e88f0d48be0899bfffacf906019bf7e31daa7d4a9c65526c24757746"} Sep 30 09:19:53 crc kubenswrapper[4810]: I0930 09:19:53.181191 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxj9j" Sep 30 09:19:53 crc kubenswrapper[4810]: I0930 09:19:53.181544 4810 scope.go:117] "RemoveContainer" containerID="fe4c9acc4cc06939f7c9800f708977d75dd88605c76a4f544000a56b174635bf" Sep 30 09:19:53 crc kubenswrapper[4810]: I0930 09:19:53.222654 4810 scope.go:117] "RemoveContainer" containerID="7f12475ee4f81c4467fecc2eab96e6b2b2c38b24a4d34aaf1368b789092c00eb" Sep 30 09:19:53 crc kubenswrapper[4810]: I0930 09:19:53.230639 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mxj9j"] Sep 30 09:19:53 crc kubenswrapper[4810]: I0930 09:19:53.244320 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mxj9j"] Sep 30 09:19:53 crc kubenswrapper[4810]: I0930 09:19:53.261635 4810 scope.go:117] "RemoveContainer" containerID="32ed01cbcf727397fa78858435db66d329e9c9c872236ef5190018c3b358cca7" Sep 30 09:19:53 crc kubenswrapper[4810]: I0930 09:19:53.321789 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31c34400-0ef5-4999-b235-582857bc9f5d" path="/var/lib/kubelet/pods/31c34400-0ef5-4999-b235-582857bc9f5d/volumes" Sep 30 09:19:54 crc kubenswrapper[4810]: I0930 09:19:54.306885 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:19:54 crc kubenswrapper[4810]: E0930 09:19:54.307683 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:20:07 crc kubenswrapper[4810]: I0930 09:20:07.313968 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:20:07 crc kubenswrapper[4810]: E0930 09:20:07.314719 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:20:20 crc kubenswrapper[4810]: I0930 09:20:20.306451 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:20:20 crc kubenswrapper[4810]: E0930 09:20:20.307372 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:20:35 crc kubenswrapper[4810]: I0930 09:20:35.306565 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:20:35 crc kubenswrapper[4810]: E0930 09:20:35.307200 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:20:48 crc kubenswrapper[4810]: I0930 09:20:48.307231 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:20:48 crc kubenswrapper[4810]: E0930 09:20:48.308647 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:20:59 crc kubenswrapper[4810]: I0930 09:20:59.310648 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:20:59 crc kubenswrapper[4810]: E0930 09:20:59.313190 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:21:13 crc kubenswrapper[4810]: I0930 09:21:13.306761 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:21:13 crc kubenswrapper[4810]: E0930 09:21:13.307627 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:21:26 crc kubenswrapper[4810]: I0930 09:21:26.307192 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:21:26 crc kubenswrapper[4810]: E0930 09:21:26.307945 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:21:40 crc kubenswrapper[4810]: I0930 09:21:40.306654 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:21:40 crc kubenswrapper[4810]: E0930 09:21:40.307558 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:21:53 crc kubenswrapper[4810]: I0930 09:21:53.306289 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:21:53 crc kubenswrapper[4810]: E0930 09:21:53.307121 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:22:04 crc kubenswrapper[4810]: I0930 09:22:04.306475 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:22:04 crc kubenswrapper[4810]: E0930 09:22:04.307258 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:22:19 crc kubenswrapper[4810]: I0930 09:22:19.307438 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:22:19 crc kubenswrapper[4810]: E0930 09:22:19.308617 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:22:24 crc kubenswrapper[4810]: E0930 09:22:24.318918 4810 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.203:37692->38.102.83.203:42869: read tcp 38.102.83.203:37692->38.102.83.203:42869: read: connection reset by peer Sep 30 09:22:30 crc kubenswrapper[4810]: E0930 09:22:30.299990 4810 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.203:36990->38.102.83.203:42869: write tcp 38.102.83.203:36990->38.102.83.203:42869: write: broken pipe Sep 30 09:22:32 crc kubenswrapper[4810]: I0930 09:22:32.307405 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:22:32 crc kubenswrapper[4810]: E0930 09:22:32.308010 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:22:44 crc kubenswrapper[4810]: I0930 09:22:44.306791 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:22:44 crc kubenswrapper[4810]: E0930 09:22:44.307407 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:22:56 crc kubenswrapper[4810]: I0930 09:22:56.307606 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:22:56 crc kubenswrapper[4810]: E0930 09:22:56.308447 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.192623 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2gj9r"] Sep 30 09:23:01 crc kubenswrapper[4810]: E0930 09:23:01.193613 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31c34400-0ef5-4999-b235-582857bc9f5d" containerName="registry-server" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.193626 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="31c34400-0ef5-4999-b235-582857bc9f5d" containerName="registry-server" Sep 30 09:23:01 crc kubenswrapper[4810]: E0930 09:23:01.193650 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="300b645a-3bec-4061-8c88-4d5cd0f2496f" containerName="extract-utilities" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.193656 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="300b645a-3bec-4061-8c88-4d5cd0f2496f" containerName="extract-utilities" Sep 30 09:23:01 crc kubenswrapper[4810]: E0930 09:23:01.193669 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31c34400-0ef5-4999-b235-582857bc9f5d" containerName="extract-utilities" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.193675 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="31c34400-0ef5-4999-b235-582857bc9f5d" containerName="extract-utilities" Sep 30 09:23:01 crc kubenswrapper[4810]: E0930 09:23:01.193689 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="300b645a-3bec-4061-8c88-4d5cd0f2496f" containerName="extract-content" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.193696 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="300b645a-3bec-4061-8c88-4d5cd0f2496f" containerName="extract-content" Sep 30 09:23:01 crc kubenswrapper[4810]: E0930 09:23:01.193710 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="300b645a-3bec-4061-8c88-4d5cd0f2496f" containerName="registry-server" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.193718 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="300b645a-3bec-4061-8c88-4d5cd0f2496f" containerName="registry-server" Sep 30 09:23:01 crc kubenswrapper[4810]: E0930 09:23:01.193745 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31c34400-0ef5-4999-b235-582857bc9f5d" containerName="extract-content" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.193752 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="31c34400-0ef5-4999-b235-582857bc9f5d" containerName="extract-content" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.193989 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="31c34400-0ef5-4999-b235-582857bc9f5d" containerName="registry-server" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.194023 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="300b645a-3bec-4061-8c88-4d5cd0f2496f" containerName="registry-server" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.210073 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2gj9r"] Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.210221 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.322137 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-catalog-content\") pod \"community-operators-2gj9r\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.322447 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr96r\" (UniqueName: \"kubernetes.io/projected/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-kube-api-access-pr96r\") pod \"community-operators-2gj9r\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.322525 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-utilities\") pod \"community-operators-2gj9r\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.423884 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-catalog-content\") pod \"community-operators-2gj9r\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.423974 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr96r\" (UniqueName: \"kubernetes.io/projected/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-kube-api-access-pr96r\") pod \"community-operators-2gj9r\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.423995 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-utilities\") pod \"community-operators-2gj9r\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.425115 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-catalog-content\") pod \"community-operators-2gj9r\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.425169 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-utilities\") pod \"community-operators-2gj9r\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.446906 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr96r\" (UniqueName: \"kubernetes.io/projected/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-kube-api-access-pr96r\") pod \"community-operators-2gj9r\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:01 crc kubenswrapper[4810]: I0930 09:23:01.535114 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:02 crc kubenswrapper[4810]: I0930 09:23:02.067957 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2gj9r"] Sep 30 09:23:02 crc kubenswrapper[4810]: W0930 09:23:02.806840 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00a1c1e4_9328_4ec8_9a1b_19031b0ff2e9.slice/crio-c6ca53a469beee95002bb5f55b0fc014aecdadfb8d3950c4bb73cae562612a14 WatchSource:0}: Error finding container c6ca53a469beee95002bb5f55b0fc014aecdadfb8d3950c4bb73cae562612a14: Status 404 returned error can't find the container with id c6ca53a469beee95002bb5f55b0fc014aecdadfb8d3950c4bb73cae562612a14 Sep 30 09:23:03 crc kubenswrapper[4810]: I0930 09:23:03.344148 4810 generic.go:334] "Generic (PLEG): container finished" podID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" containerID="6295d6533cef517c14a8f64962efb8e1749d81bb53ef435ac5ae993cc19ddca5" exitCode=0 Sep 30 09:23:03 crc kubenswrapper[4810]: I0930 09:23:03.344277 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gj9r" event={"ID":"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9","Type":"ContainerDied","Data":"6295d6533cef517c14a8f64962efb8e1749d81bb53ef435ac5ae993cc19ddca5"} Sep 30 09:23:03 crc kubenswrapper[4810]: I0930 09:23:03.344528 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gj9r" event={"ID":"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9","Type":"ContainerStarted","Data":"c6ca53a469beee95002bb5f55b0fc014aecdadfb8d3950c4bb73cae562612a14"} Sep 30 09:23:03 crc kubenswrapper[4810]: I0930 09:23:03.346139 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 09:23:05 crc kubenswrapper[4810]: I0930 09:23:05.366696 4810 generic.go:334] "Generic (PLEG): container finished" podID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" containerID="44cab7922b3b9da3b93a515c202d858823a0576d8dbb513a1ca7f8a5771928c2" exitCode=0 Sep 30 09:23:05 crc kubenswrapper[4810]: I0930 09:23:05.366762 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gj9r" event={"ID":"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9","Type":"ContainerDied","Data":"44cab7922b3b9da3b93a515c202d858823a0576d8dbb513a1ca7f8a5771928c2"} Sep 30 09:23:06 crc kubenswrapper[4810]: I0930 09:23:06.379259 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gj9r" event={"ID":"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9","Type":"ContainerStarted","Data":"dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b"} Sep 30 09:23:06 crc kubenswrapper[4810]: I0930 09:23:06.399733 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2gj9r" podStartSLOduration=2.9537598000000003 podStartE2EDuration="5.399695358s" podCreationTimestamp="2025-09-30 09:23:01 +0000 UTC" firstStartedPulling="2025-09-30 09:23:03.345873017 +0000 UTC m=+4806.798072294" lastFinishedPulling="2025-09-30 09:23:05.791808535 +0000 UTC m=+4809.244007852" observedRunningTime="2025-09-30 09:23:06.395242909 +0000 UTC m=+4809.847442176" watchObservedRunningTime="2025-09-30 09:23:06.399695358 +0000 UTC m=+4809.851894625" Sep 30 09:23:10 crc kubenswrapper[4810]: I0930 09:23:10.306672 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:23:10 crc kubenswrapper[4810]: E0930 09:23:10.307646 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:23:11 crc kubenswrapper[4810]: I0930 09:23:11.535701 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:11 crc kubenswrapper[4810]: I0930 09:23:11.536506 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:11 crc kubenswrapper[4810]: I0930 09:23:11.588619 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:12 crc kubenswrapper[4810]: I0930 09:23:12.514035 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:12 crc kubenswrapper[4810]: I0930 09:23:12.559234 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2gj9r"] Sep 30 09:23:14 crc kubenswrapper[4810]: I0930 09:23:14.480011 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2gj9r" podUID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" containerName="registry-server" containerID="cri-o://dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b" gracePeriod=2 Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.068577 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.118808 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr96r\" (UniqueName: \"kubernetes.io/projected/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-kube-api-access-pr96r\") pod \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.119184 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-utilities\") pod \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.119391 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-catalog-content\") pod \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\" (UID: \"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9\") " Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.120483 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-utilities" (OuterVolumeSpecName: "utilities") pod "00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" (UID: "00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.128258 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-kube-api-access-pr96r" (OuterVolumeSpecName: "kube-api-access-pr96r") pod "00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" (UID: "00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9"). InnerVolumeSpecName "kube-api-access-pr96r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.201834 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" (UID: "00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.222540 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr96r\" (UniqueName: \"kubernetes.io/projected/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-kube-api-access-pr96r\") on node \"crc\" DevicePath \"\"" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.222570 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.222583 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.492939 4810 generic.go:334] "Generic (PLEG): container finished" podID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" containerID="dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b" exitCode=0 Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.493003 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gj9r" event={"ID":"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9","Type":"ContainerDied","Data":"dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b"} Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.493053 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gj9r" event={"ID":"00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9","Type":"ContainerDied","Data":"c6ca53a469beee95002bb5f55b0fc014aecdadfb8d3950c4bb73cae562612a14"} Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.493077 4810 scope.go:117] "RemoveContainer" containerID="dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.493017 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gj9r" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.516974 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2gj9r"] Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.524206 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2gj9r"] Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.535803 4810 scope.go:117] "RemoveContainer" containerID="44cab7922b3b9da3b93a515c202d858823a0576d8dbb513a1ca7f8a5771928c2" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.558126 4810 scope.go:117] "RemoveContainer" containerID="6295d6533cef517c14a8f64962efb8e1749d81bb53ef435ac5ae993cc19ddca5" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.608469 4810 scope.go:117] "RemoveContainer" containerID="dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b" Sep 30 09:23:15 crc kubenswrapper[4810]: E0930 09:23:15.609059 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b\": container with ID starting with dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b not found: ID does not exist" containerID="dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.609096 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b"} err="failed to get container status \"dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b\": rpc error: code = NotFound desc = could not find container \"dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b\": container with ID starting with dd1d1beb475ac15582a2d53d146e1d400939cafb5abf5bce085cbeb3b4e8931b not found: ID does not exist" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.609122 4810 scope.go:117] "RemoveContainer" containerID="44cab7922b3b9da3b93a515c202d858823a0576d8dbb513a1ca7f8a5771928c2" Sep 30 09:23:15 crc kubenswrapper[4810]: E0930 09:23:15.609411 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44cab7922b3b9da3b93a515c202d858823a0576d8dbb513a1ca7f8a5771928c2\": container with ID starting with 44cab7922b3b9da3b93a515c202d858823a0576d8dbb513a1ca7f8a5771928c2 not found: ID does not exist" containerID="44cab7922b3b9da3b93a515c202d858823a0576d8dbb513a1ca7f8a5771928c2" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.609439 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44cab7922b3b9da3b93a515c202d858823a0576d8dbb513a1ca7f8a5771928c2"} err="failed to get container status \"44cab7922b3b9da3b93a515c202d858823a0576d8dbb513a1ca7f8a5771928c2\": rpc error: code = NotFound desc = could not find container \"44cab7922b3b9da3b93a515c202d858823a0576d8dbb513a1ca7f8a5771928c2\": container with ID starting with 44cab7922b3b9da3b93a515c202d858823a0576d8dbb513a1ca7f8a5771928c2 not found: ID does not exist" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.609452 4810 scope.go:117] "RemoveContainer" containerID="6295d6533cef517c14a8f64962efb8e1749d81bb53ef435ac5ae993cc19ddca5" Sep 30 09:23:15 crc kubenswrapper[4810]: E0930 09:23:15.609758 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6295d6533cef517c14a8f64962efb8e1749d81bb53ef435ac5ae993cc19ddca5\": container with ID starting with 6295d6533cef517c14a8f64962efb8e1749d81bb53ef435ac5ae993cc19ddca5 not found: ID does not exist" containerID="6295d6533cef517c14a8f64962efb8e1749d81bb53ef435ac5ae993cc19ddca5" Sep 30 09:23:15 crc kubenswrapper[4810]: I0930 09:23:15.609785 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6295d6533cef517c14a8f64962efb8e1749d81bb53ef435ac5ae993cc19ddca5"} err="failed to get container status \"6295d6533cef517c14a8f64962efb8e1749d81bb53ef435ac5ae993cc19ddca5\": rpc error: code = NotFound desc = could not find container \"6295d6533cef517c14a8f64962efb8e1749d81bb53ef435ac5ae993cc19ddca5\": container with ID starting with 6295d6533cef517c14a8f64962efb8e1749d81bb53ef435ac5ae993cc19ddca5 not found: ID does not exist" Sep 30 09:23:17 crc kubenswrapper[4810]: I0930 09:23:17.321603 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" path="/var/lib/kubelet/pods/00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9/volumes" Sep 30 09:23:22 crc kubenswrapper[4810]: I0930 09:23:22.307779 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:23:22 crc kubenswrapper[4810]: E0930 09:23:22.310851 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:23:37 crc kubenswrapper[4810]: I0930 09:23:37.313540 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:23:37 crc kubenswrapper[4810]: E0930 09:23:37.314499 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:23:51 crc kubenswrapper[4810]: I0930 09:23:51.308354 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:23:51 crc kubenswrapper[4810]: E0930 09:23:51.309395 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:24:04 crc kubenswrapper[4810]: I0930 09:24:04.306029 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:24:04 crc kubenswrapper[4810]: E0930 09:24:04.306775 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:24:18 crc kubenswrapper[4810]: I0930 09:24:18.306943 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:24:19 crc kubenswrapper[4810]: I0930 09:24:19.189089 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"fdefe6392dc2bfa1f689117fddb7ef6ba2f43e8dcedf0924a7ce3f58a2613787"} Sep 30 09:26:45 crc kubenswrapper[4810]: I0930 09:26:45.911854 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:26:45 crc kubenswrapper[4810]: I0930 09:26:45.912730 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:27:15 crc kubenswrapper[4810]: I0930 09:27:15.911116 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:27:15 crc kubenswrapper[4810]: I0930 09:27:15.911891 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:27:45 crc kubenswrapper[4810]: I0930 09:27:45.911866 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:27:45 crc kubenswrapper[4810]: I0930 09:27:45.913559 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:27:45 crc kubenswrapper[4810]: I0930 09:27:45.913652 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 09:27:45 crc kubenswrapper[4810]: I0930 09:27:45.914737 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fdefe6392dc2bfa1f689117fddb7ef6ba2f43e8dcedf0924a7ce3f58a2613787"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 09:27:45 crc kubenswrapper[4810]: I0930 09:27:45.914847 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://fdefe6392dc2bfa1f689117fddb7ef6ba2f43e8dcedf0924a7ce3f58a2613787" gracePeriod=600 Sep 30 09:27:46 crc kubenswrapper[4810]: I0930 09:27:46.556705 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="fdefe6392dc2bfa1f689117fddb7ef6ba2f43e8dcedf0924a7ce3f58a2613787" exitCode=0 Sep 30 09:27:46 crc kubenswrapper[4810]: I0930 09:27:46.557377 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"fdefe6392dc2bfa1f689117fddb7ef6ba2f43e8dcedf0924a7ce3f58a2613787"} Sep 30 09:27:46 crc kubenswrapper[4810]: I0930 09:27:46.557587 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68"} Sep 30 09:27:46 crc kubenswrapper[4810]: I0930 09:27:46.557618 4810 scope.go:117] "RemoveContainer" containerID="b7b0b41d9367d7718327c35b8055e790f9da0533321d66ea3dd809575c00f01f" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.646942 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2xktw"] Sep 30 09:28:45 crc kubenswrapper[4810]: E0930 09:28:45.648068 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" containerName="extract-content" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.648086 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" containerName="extract-content" Sep 30 09:28:45 crc kubenswrapper[4810]: E0930 09:28:45.648115 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" containerName="registry-server" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.648125 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" containerName="registry-server" Sep 30 09:28:45 crc kubenswrapper[4810]: E0930 09:28:45.648185 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" containerName="extract-utilities" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.648196 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" containerName="extract-utilities" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.648475 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a1c1e4-9328-4ec8-9a1b-19031b0ff2e9" containerName="registry-server" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.653476 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.665988 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2xktw"] Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.737668 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-catalog-content\") pod \"redhat-marketplace-2xktw\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.738047 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhwrm\" (UniqueName: \"kubernetes.io/projected/9013a27f-a30f-4d87-b23d-8311ea353beb-kube-api-access-vhwrm\") pod \"redhat-marketplace-2xktw\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.738250 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-utilities\") pod \"redhat-marketplace-2xktw\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.840248 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-catalog-content\") pod \"redhat-marketplace-2xktw\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.840477 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhwrm\" (UniqueName: \"kubernetes.io/projected/9013a27f-a30f-4d87-b23d-8311ea353beb-kube-api-access-vhwrm\") pod \"redhat-marketplace-2xktw\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.840550 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-utilities\") pod \"redhat-marketplace-2xktw\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.841108 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-catalog-content\") pod \"redhat-marketplace-2xktw\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.841377 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-utilities\") pod \"redhat-marketplace-2xktw\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.863624 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhwrm\" (UniqueName: \"kubernetes.io/projected/9013a27f-a30f-4d87-b23d-8311ea353beb-kube-api-access-vhwrm\") pod \"redhat-marketplace-2xktw\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:45 crc kubenswrapper[4810]: I0930 09:28:45.993481 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:46 crc kubenswrapper[4810]: I0930 09:28:46.582244 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2xktw"] Sep 30 09:28:47 crc kubenswrapper[4810]: I0930 09:28:47.266345 4810 generic.go:334] "Generic (PLEG): container finished" podID="9013a27f-a30f-4d87-b23d-8311ea353beb" containerID="2ad8b9b7ec6b8ee7fe73f1a17b9f5b0fccd94d6be7a62e446908c6d34c38f257" exitCode=0 Sep 30 09:28:47 crc kubenswrapper[4810]: I0930 09:28:47.266413 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xktw" event={"ID":"9013a27f-a30f-4d87-b23d-8311ea353beb","Type":"ContainerDied","Data":"2ad8b9b7ec6b8ee7fe73f1a17b9f5b0fccd94d6be7a62e446908c6d34c38f257"} Sep 30 09:28:47 crc kubenswrapper[4810]: I0930 09:28:47.266445 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xktw" event={"ID":"9013a27f-a30f-4d87-b23d-8311ea353beb","Type":"ContainerStarted","Data":"688fd44a82bd694be1bc9f61575f9c8a633b39100614ddb2a5a5efc84dbea6a5"} Sep 30 09:28:47 crc kubenswrapper[4810]: I0930 09:28:47.270364 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 09:28:48 crc kubenswrapper[4810]: I0930 09:28:48.281819 4810 generic.go:334] "Generic (PLEG): container finished" podID="9013a27f-a30f-4d87-b23d-8311ea353beb" containerID="d5e69ef35b6c371cfd1da2dd6a3ff34f969655442734b0b96b6d074d0bebc090" exitCode=0 Sep 30 09:28:48 crc kubenswrapper[4810]: I0930 09:28:48.281900 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xktw" event={"ID":"9013a27f-a30f-4d87-b23d-8311ea353beb","Type":"ContainerDied","Data":"d5e69ef35b6c371cfd1da2dd6a3ff34f969655442734b0b96b6d074d0bebc090"} Sep 30 09:28:49 crc kubenswrapper[4810]: I0930 09:28:49.297022 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xktw" event={"ID":"9013a27f-a30f-4d87-b23d-8311ea353beb","Type":"ContainerStarted","Data":"d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d"} Sep 30 09:28:49 crc kubenswrapper[4810]: I0930 09:28:49.327046 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2xktw" podStartSLOduration=2.865701773 podStartE2EDuration="4.327017655s" podCreationTimestamp="2025-09-30 09:28:45 +0000 UTC" firstStartedPulling="2025-09-30 09:28:47.270116581 +0000 UTC m=+5150.722315848" lastFinishedPulling="2025-09-30 09:28:48.731432463 +0000 UTC m=+5152.183631730" observedRunningTime="2025-09-30 09:28:49.318579771 +0000 UTC m=+5152.770779058" watchObservedRunningTime="2025-09-30 09:28:49.327017655 +0000 UTC m=+5152.779216962" Sep 30 09:28:55 crc kubenswrapper[4810]: I0930 09:28:55.994545 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:55 crc kubenswrapper[4810]: I0930 09:28:55.996374 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:56 crc kubenswrapper[4810]: I0930 09:28:56.060768 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:56 crc kubenswrapper[4810]: I0930 09:28:56.424481 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:56 crc kubenswrapper[4810]: I0930 09:28:56.491518 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2xktw"] Sep 30 09:28:58 crc kubenswrapper[4810]: I0930 09:28:58.378710 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2xktw" podUID="9013a27f-a30f-4d87-b23d-8311ea353beb" containerName="registry-server" containerID="cri-o://d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d" gracePeriod=2 Sep 30 09:28:58 crc kubenswrapper[4810]: I0930 09:28:58.872867 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.032212 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-utilities\") pod \"9013a27f-a30f-4d87-b23d-8311ea353beb\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.032356 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-catalog-content\") pod \"9013a27f-a30f-4d87-b23d-8311ea353beb\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.032457 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhwrm\" (UniqueName: \"kubernetes.io/projected/9013a27f-a30f-4d87-b23d-8311ea353beb-kube-api-access-vhwrm\") pod \"9013a27f-a30f-4d87-b23d-8311ea353beb\" (UID: \"9013a27f-a30f-4d87-b23d-8311ea353beb\") " Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.033172 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-utilities" (OuterVolumeSpecName: "utilities") pod "9013a27f-a30f-4d87-b23d-8311ea353beb" (UID: "9013a27f-a30f-4d87-b23d-8311ea353beb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.033457 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.039288 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9013a27f-a30f-4d87-b23d-8311ea353beb-kube-api-access-vhwrm" (OuterVolumeSpecName: "kube-api-access-vhwrm") pod "9013a27f-a30f-4d87-b23d-8311ea353beb" (UID: "9013a27f-a30f-4d87-b23d-8311ea353beb"). InnerVolumeSpecName "kube-api-access-vhwrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.067549 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9013a27f-a30f-4d87-b23d-8311ea353beb" (UID: "9013a27f-a30f-4d87-b23d-8311ea353beb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.135936 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9013a27f-a30f-4d87-b23d-8311ea353beb-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.135989 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhwrm\" (UniqueName: \"kubernetes.io/projected/9013a27f-a30f-4d87-b23d-8311ea353beb-kube-api-access-vhwrm\") on node \"crc\" DevicePath \"\"" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.388811 4810 generic.go:334] "Generic (PLEG): container finished" podID="9013a27f-a30f-4d87-b23d-8311ea353beb" containerID="d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d" exitCode=0 Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.388882 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2xktw" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.388910 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xktw" event={"ID":"9013a27f-a30f-4d87-b23d-8311ea353beb","Type":"ContainerDied","Data":"d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d"} Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.389222 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2xktw" event={"ID":"9013a27f-a30f-4d87-b23d-8311ea353beb","Type":"ContainerDied","Data":"688fd44a82bd694be1bc9f61575f9c8a633b39100614ddb2a5a5efc84dbea6a5"} Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.389253 4810 scope.go:117] "RemoveContainer" containerID="d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.415644 4810 scope.go:117] "RemoveContainer" containerID="d5e69ef35b6c371cfd1da2dd6a3ff34f969655442734b0b96b6d074d0bebc090" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.424044 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2xktw"] Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.433672 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2xktw"] Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.447310 4810 scope.go:117] "RemoveContainer" containerID="2ad8b9b7ec6b8ee7fe73f1a17b9f5b0fccd94d6be7a62e446908c6d34c38f257" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.512876 4810 scope.go:117] "RemoveContainer" containerID="d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d" Sep 30 09:28:59 crc kubenswrapper[4810]: E0930 09:28:59.513404 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d\": container with ID starting with d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d not found: ID does not exist" containerID="d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.513450 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d"} err="failed to get container status \"d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d\": rpc error: code = NotFound desc = could not find container \"d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d\": container with ID starting with d590d5f5501f68770b370e028d293579cca36233fc3b6b1233c5d1901c799d3d not found: ID does not exist" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.513486 4810 scope.go:117] "RemoveContainer" containerID="d5e69ef35b6c371cfd1da2dd6a3ff34f969655442734b0b96b6d074d0bebc090" Sep 30 09:28:59 crc kubenswrapper[4810]: E0930 09:28:59.513967 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5e69ef35b6c371cfd1da2dd6a3ff34f969655442734b0b96b6d074d0bebc090\": container with ID starting with d5e69ef35b6c371cfd1da2dd6a3ff34f969655442734b0b96b6d074d0bebc090 not found: ID does not exist" containerID="d5e69ef35b6c371cfd1da2dd6a3ff34f969655442734b0b96b6d074d0bebc090" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.514046 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5e69ef35b6c371cfd1da2dd6a3ff34f969655442734b0b96b6d074d0bebc090"} err="failed to get container status \"d5e69ef35b6c371cfd1da2dd6a3ff34f969655442734b0b96b6d074d0bebc090\": rpc error: code = NotFound desc = could not find container \"d5e69ef35b6c371cfd1da2dd6a3ff34f969655442734b0b96b6d074d0bebc090\": container with ID starting with d5e69ef35b6c371cfd1da2dd6a3ff34f969655442734b0b96b6d074d0bebc090 not found: ID does not exist" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.514092 4810 scope.go:117] "RemoveContainer" containerID="2ad8b9b7ec6b8ee7fe73f1a17b9f5b0fccd94d6be7a62e446908c6d34c38f257" Sep 30 09:28:59 crc kubenswrapper[4810]: E0930 09:28:59.514468 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ad8b9b7ec6b8ee7fe73f1a17b9f5b0fccd94d6be7a62e446908c6d34c38f257\": container with ID starting with 2ad8b9b7ec6b8ee7fe73f1a17b9f5b0fccd94d6be7a62e446908c6d34c38f257 not found: ID does not exist" containerID="2ad8b9b7ec6b8ee7fe73f1a17b9f5b0fccd94d6be7a62e446908c6d34c38f257" Sep 30 09:28:59 crc kubenswrapper[4810]: I0930 09:28:59.514519 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ad8b9b7ec6b8ee7fe73f1a17b9f5b0fccd94d6be7a62e446908c6d34c38f257"} err="failed to get container status \"2ad8b9b7ec6b8ee7fe73f1a17b9f5b0fccd94d6be7a62e446908c6d34c38f257\": rpc error: code = NotFound desc = could not find container \"2ad8b9b7ec6b8ee7fe73f1a17b9f5b0fccd94d6be7a62e446908c6d34c38f257\": container with ID starting with 2ad8b9b7ec6b8ee7fe73f1a17b9f5b0fccd94d6be7a62e446908c6d34c38f257 not found: ID does not exist" Sep 30 09:29:01 crc kubenswrapper[4810]: I0930 09:29:01.317379 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9013a27f-a30f-4d87-b23d-8311ea353beb" path="/var/lib/kubelet/pods/9013a27f-a30f-4d87-b23d-8311ea353beb/volumes" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.156531 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq"] Sep 30 09:30:00 crc kubenswrapper[4810]: E0930 09:30:00.157505 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9013a27f-a30f-4d87-b23d-8311ea353beb" containerName="extract-utilities" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.157521 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9013a27f-a30f-4d87-b23d-8311ea353beb" containerName="extract-utilities" Sep 30 09:30:00 crc kubenswrapper[4810]: E0930 09:30:00.157540 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9013a27f-a30f-4d87-b23d-8311ea353beb" containerName="registry-server" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.157547 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9013a27f-a30f-4d87-b23d-8311ea353beb" containerName="registry-server" Sep 30 09:30:00 crc kubenswrapper[4810]: E0930 09:30:00.157587 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9013a27f-a30f-4d87-b23d-8311ea353beb" containerName="extract-content" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.157595 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9013a27f-a30f-4d87-b23d-8311ea353beb" containerName="extract-content" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.157831 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9013a27f-a30f-4d87-b23d-8311ea353beb" containerName="registry-server" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.158701 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.162290 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.162377 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.172598 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq"] Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.210390 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-secret-volume\") pod \"collect-profiles-29320410-dmmhq\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.210483 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45krn\" (UniqueName: \"kubernetes.io/projected/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-kube-api-access-45krn\") pod \"collect-profiles-29320410-dmmhq\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.210674 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-config-volume\") pod \"collect-profiles-29320410-dmmhq\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.312492 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45krn\" (UniqueName: \"kubernetes.io/projected/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-kube-api-access-45krn\") pod \"collect-profiles-29320410-dmmhq\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.312658 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-config-volume\") pod \"collect-profiles-29320410-dmmhq\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.312722 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-secret-volume\") pod \"collect-profiles-29320410-dmmhq\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.313416 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-config-volume\") pod \"collect-profiles-29320410-dmmhq\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.318045 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-secret-volume\") pod \"collect-profiles-29320410-dmmhq\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.332860 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45krn\" (UniqueName: \"kubernetes.io/projected/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-kube-api-access-45krn\") pod \"collect-profiles-29320410-dmmhq\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.486834 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:00 crc kubenswrapper[4810]: I0930 09:30:00.925250 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq"] Sep 30 09:30:00 crc kubenswrapper[4810]: W0930 09:30:00.933582 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07f4c77f_1ac0_47eb_9f6f_42ffaa34036e.slice/crio-e50201a18343097342740d7229d3536f9b66245b3fcd48b4353d4a38745e6f3d WatchSource:0}: Error finding container e50201a18343097342740d7229d3536f9b66245b3fcd48b4353d4a38745e6f3d: Status 404 returned error can't find the container with id e50201a18343097342740d7229d3536f9b66245b3fcd48b4353d4a38745e6f3d Sep 30 09:30:01 crc kubenswrapper[4810]: I0930 09:30:01.072245 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" event={"ID":"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e","Type":"ContainerStarted","Data":"e50201a18343097342740d7229d3536f9b66245b3fcd48b4353d4a38745e6f3d"} Sep 30 09:30:02 crc kubenswrapper[4810]: I0930 09:30:02.088128 4810 generic.go:334] "Generic (PLEG): container finished" podID="07f4c77f-1ac0-47eb-9f6f-42ffaa34036e" containerID="6ab57ed4b3495b24cb6ed9480a009c4376daf0d6393a437e1002ec4c76d5166d" exitCode=0 Sep 30 09:30:02 crc kubenswrapper[4810]: I0930 09:30:02.088235 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" event={"ID":"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e","Type":"ContainerDied","Data":"6ab57ed4b3495b24cb6ed9480a009c4376daf0d6393a437e1002ec4c76d5166d"} Sep 30 09:30:03 crc kubenswrapper[4810]: I0930 09:30:03.483363 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:03 crc kubenswrapper[4810]: I0930 09:30:03.579559 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45krn\" (UniqueName: \"kubernetes.io/projected/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-kube-api-access-45krn\") pod \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " Sep 30 09:30:03 crc kubenswrapper[4810]: I0930 09:30:03.579616 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-config-volume\") pod \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " Sep 30 09:30:03 crc kubenswrapper[4810]: I0930 09:30:03.579704 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-secret-volume\") pod \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\" (UID: \"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e\") " Sep 30 09:30:03 crc kubenswrapper[4810]: I0930 09:30:03.580696 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-config-volume" (OuterVolumeSpecName: "config-volume") pod "07f4c77f-1ac0-47eb-9f6f-42ffaa34036e" (UID: "07f4c77f-1ac0-47eb-9f6f-42ffaa34036e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 09:30:03 crc kubenswrapper[4810]: I0930 09:30:03.589910 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "07f4c77f-1ac0-47eb-9f6f-42ffaa34036e" (UID: "07f4c77f-1ac0-47eb-9f6f-42ffaa34036e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 09:30:03 crc kubenswrapper[4810]: I0930 09:30:03.590491 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-kube-api-access-45krn" (OuterVolumeSpecName: "kube-api-access-45krn") pod "07f4c77f-1ac0-47eb-9f6f-42ffaa34036e" (UID: "07f4c77f-1ac0-47eb-9f6f-42ffaa34036e"). InnerVolumeSpecName "kube-api-access-45krn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:30:03 crc kubenswrapper[4810]: I0930 09:30:03.682347 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 09:30:03 crc kubenswrapper[4810]: I0930 09:30:03.682397 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 09:30:03 crc kubenswrapper[4810]: I0930 09:30:03.682410 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45krn\" (UniqueName: \"kubernetes.io/projected/07f4c77f-1ac0-47eb-9f6f-42ffaa34036e-kube-api-access-45krn\") on node \"crc\" DevicePath \"\"" Sep 30 09:30:04 crc kubenswrapper[4810]: I0930 09:30:04.121393 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" event={"ID":"07f4c77f-1ac0-47eb-9f6f-42ffaa34036e","Type":"ContainerDied","Data":"e50201a18343097342740d7229d3536f9b66245b3fcd48b4353d4a38745e6f3d"} Sep 30 09:30:04 crc kubenswrapper[4810]: I0930 09:30:04.121467 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e50201a18343097342740d7229d3536f9b66245b3fcd48b4353d4a38745e6f3d" Sep 30 09:30:04 crc kubenswrapper[4810]: I0930 09:30:04.121560 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320410-dmmhq" Sep 30 09:30:04 crc kubenswrapper[4810]: I0930 09:30:04.552960 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s"] Sep 30 09:30:04 crc kubenswrapper[4810]: I0930 09:30:04.561729 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320365-q545s"] Sep 30 09:30:05 crc kubenswrapper[4810]: I0930 09:30:05.320486 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07460fac-bba2-429f-b0f0-5e569719e887" path="/var/lib/kubelet/pods/07460fac-bba2-429f-b0f0-5e569719e887/volumes" Sep 30 09:30:15 crc kubenswrapper[4810]: I0930 09:30:15.911806 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:30:15 crc kubenswrapper[4810]: I0930 09:30:15.913221 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:30:23 crc kubenswrapper[4810]: I0930 09:30:23.199781 4810 scope.go:117] "RemoveContainer" containerID="3aaea24f70ebaa8415966583515ecaa5dc71ef91585138b665cafe07b2c4396b" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.758854 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fxchk"] Sep 30 09:30:25 crc kubenswrapper[4810]: E0930 09:30:25.760080 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07f4c77f-1ac0-47eb-9f6f-42ffaa34036e" containerName="collect-profiles" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.760101 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="07f4c77f-1ac0-47eb-9f6f-42ffaa34036e" containerName="collect-profiles" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.760468 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="07f4c77f-1ac0-47eb-9f6f-42ffaa34036e" containerName="collect-profiles" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.762930 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.781406 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fxchk"] Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.871556 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt8rc\" (UniqueName: \"kubernetes.io/projected/72e49d38-ec21-403d-9f29-97a3b9415cab-kube-api-access-nt8rc\") pod \"certified-operators-fxchk\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.871653 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-utilities\") pod \"certified-operators-fxchk\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.871777 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-catalog-content\") pod \"certified-operators-fxchk\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.973677 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-catalog-content\") pod \"certified-operators-fxchk\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.974000 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt8rc\" (UniqueName: \"kubernetes.io/projected/72e49d38-ec21-403d-9f29-97a3b9415cab-kube-api-access-nt8rc\") pod \"certified-operators-fxchk\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.974111 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-utilities\") pod \"certified-operators-fxchk\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.974340 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-catalog-content\") pod \"certified-operators-fxchk\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:25 crc kubenswrapper[4810]: I0930 09:30:25.974527 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-utilities\") pod \"certified-operators-fxchk\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:26 crc kubenswrapper[4810]: I0930 09:30:26.089819 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt8rc\" (UniqueName: \"kubernetes.io/projected/72e49d38-ec21-403d-9f29-97a3b9415cab-kube-api-access-nt8rc\") pod \"certified-operators-fxchk\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:26 crc kubenswrapper[4810]: I0930 09:30:26.137859 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:26 crc kubenswrapper[4810]: W0930 09:30:26.752322 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72e49d38_ec21_403d_9f29_97a3b9415cab.slice/crio-4a50809cf3e4d1642d04026937bd7718002f53f54b05fdb6ea69126852b36b12 WatchSource:0}: Error finding container 4a50809cf3e4d1642d04026937bd7718002f53f54b05fdb6ea69126852b36b12: Status 404 returned error can't find the container with id 4a50809cf3e4d1642d04026937bd7718002f53f54b05fdb6ea69126852b36b12 Sep 30 09:30:26 crc kubenswrapper[4810]: I0930 09:30:26.760567 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fxchk"] Sep 30 09:30:27 crc kubenswrapper[4810]: I0930 09:30:27.387587 4810 generic.go:334] "Generic (PLEG): container finished" podID="72e49d38-ec21-403d-9f29-97a3b9415cab" containerID="922c6de228a2e65b09a8aab043431ebcdf7e1b38834b20279a284ae965cda435" exitCode=0 Sep 30 09:30:27 crc kubenswrapper[4810]: I0930 09:30:27.387915 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxchk" event={"ID":"72e49d38-ec21-403d-9f29-97a3b9415cab","Type":"ContainerDied","Data":"922c6de228a2e65b09a8aab043431ebcdf7e1b38834b20279a284ae965cda435"} Sep 30 09:30:27 crc kubenswrapper[4810]: I0930 09:30:27.387945 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxchk" event={"ID":"72e49d38-ec21-403d-9f29-97a3b9415cab","Type":"ContainerStarted","Data":"4a50809cf3e4d1642d04026937bd7718002f53f54b05fdb6ea69126852b36b12"} Sep 30 09:30:29 crc kubenswrapper[4810]: I0930 09:30:29.409097 4810 generic.go:334] "Generic (PLEG): container finished" podID="72e49d38-ec21-403d-9f29-97a3b9415cab" containerID="a5eb2fe712743b6df5b92327728adda9631d0629b363beac574ec4f9a9265142" exitCode=0 Sep 30 09:30:29 crc kubenswrapper[4810]: I0930 09:30:29.409743 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxchk" event={"ID":"72e49d38-ec21-403d-9f29-97a3b9415cab","Type":"ContainerDied","Data":"a5eb2fe712743b6df5b92327728adda9631d0629b363beac574ec4f9a9265142"} Sep 30 09:30:31 crc kubenswrapper[4810]: I0930 09:30:31.433377 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxchk" event={"ID":"72e49d38-ec21-403d-9f29-97a3b9415cab","Type":"ContainerStarted","Data":"0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea"} Sep 30 09:30:31 crc kubenswrapper[4810]: I0930 09:30:31.464969 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fxchk" podStartSLOduration=3.6104590500000002 podStartE2EDuration="6.464952843s" podCreationTimestamp="2025-09-30 09:30:25 +0000 UTC" firstStartedPulling="2025-09-30 09:30:27.390163041 +0000 UTC m=+5250.842362308" lastFinishedPulling="2025-09-30 09:30:30.244656844 +0000 UTC m=+5253.696856101" observedRunningTime="2025-09-30 09:30:31.462709708 +0000 UTC m=+5254.914908985" watchObservedRunningTime="2025-09-30 09:30:31.464952843 +0000 UTC m=+5254.917152130" Sep 30 09:30:36 crc kubenswrapper[4810]: I0930 09:30:36.139767 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:36 crc kubenswrapper[4810]: I0930 09:30:36.140358 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:36 crc kubenswrapper[4810]: I0930 09:30:36.203579 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:36 crc kubenswrapper[4810]: I0930 09:30:36.553342 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:36 crc kubenswrapper[4810]: I0930 09:30:36.605727 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fxchk"] Sep 30 09:30:38 crc kubenswrapper[4810]: I0930 09:30:38.513387 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fxchk" podUID="72e49d38-ec21-403d-9f29-97a3b9415cab" containerName="registry-server" containerID="cri-o://0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea" gracePeriod=2 Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.050305 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.154329 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nt8rc\" (UniqueName: \"kubernetes.io/projected/72e49d38-ec21-403d-9f29-97a3b9415cab-kube-api-access-nt8rc\") pod \"72e49d38-ec21-403d-9f29-97a3b9415cab\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.154383 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-utilities\") pod \"72e49d38-ec21-403d-9f29-97a3b9415cab\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.154421 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-catalog-content\") pod \"72e49d38-ec21-403d-9f29-97a3b9415cab\" (UID: \"72e49d38-ec21-403d-9f29-97a3b9415cab\") " Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.155674 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-utilities" (OuterVolumeSpecName: "utilities") pod "72e49d38-ec21-403d-9f29-97a3b9415cab" (UID: "72e49d38-ec21-403d-9f29-97a3b9415cab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.156346 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.198047 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72e49d38-ec21-403d-9f29-97a3b9415cab" (UID: "72e49d38-ec21-403d-9f29-97a3b9415cab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.258089 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e49d38-ec21-403d-9f29-97a3b9415cab-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.527137 4810 generic.go:334] "Generic (PLEG): container finished" podID="72e49d38-ec21-403d-9f29-97a3b9415cab" containerID="0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea" exitCode=0 Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.527216 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxchk" Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.527249 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxchk" event={"ID":"72e49d38-ec21-403d-9f29-97a3b9415cab","Type":"ContainerDied","Data":"0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea"} Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.528208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxchk" event={"ID":"72e49d38-ec21-403d-9f29-97a3b9415cab","Type":"ContainerDied","Data":"4a50809cf3e4d1642d04026937bd7718002f53f54b05fdb6ea69126852b36b12"} Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.528243 4810 scope.go:117] "RemoveContainer" containerID="0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea" Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.549125 4810 scope.go:117] "RemoveContainer" containerID="a5eb2fe712743b6df5b92327728adda9631d0629b363beac574ec4f9a9265142" Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.899012 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e49d38-ec21-403d-9f29-97a3b9415cab-kube-api-access-nt8rc" (OuterVolumeSpecName: "kube-api-access-nt8rc") pod "72e49d38-ec21-403d-9f29-97a3b9415cab" (UID: "72e49d38-ec21-403d-9f29-97a3b9415cab"). InnerVolumeSpecName "kube-api-access-nt8rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.916442 4810 scope.go:117] "RemoveContainer" containerID="922c6de228a2e65b09a8aab043431ebcdf7e1b38834b20279a284ae965cda435" Sep 30 09:30:39 crc kubenswrapper[4810]: I0930 09:30:39.974023 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nt8rc\" (UniqueName: \"kubernetes.io/projected/72e49d38-ec21-403d-9f29-97a3b9415cab-kube-api-access-nt8rc\") on node \"crc\" DevicePath \"\"" Sep 30 09:30:40 crc kubenswrapper[4810]: I0930 09:30:40.098294 4810 scope.go:117] "RemoveContainer" containerID="0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea" Sep 30 09:30:40 crc kubenswrapper[4810]: E0930 09:30:40.098740 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea\": container with ID starting with 0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea not found: ID does not exist" containerID="0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea" Sep 30 09:30:40 crc kubenswrapper[4810]: I0930 09:30:40.098773 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea"} err="failed to get container status \"0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea\": rpc error: code = NotFound desc = could not find container \"0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea\": container with ID starting with 0c4a35504b765d1f64c28132e4ba4be4e6e813f086a496ced07eaf83eeb8e4ea not found: ID does not exist" Sep 30 09:30:40 crc kubenswrapper[4810]: I0930 09:30:40.098799 4810 scope.go:117] "RemoveContainer" containerID="a5eb2fe712743b6df5b92327728adda9631d0629b363beac574ec4f9a9265142" Sep 30 09:30:40 crc kubenswrapper[4810]: E0930 09:30:40.099057 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5eb2fe712743b6df5b92327728adda9631d0629b363beac574ec4f9a9265142\": container with ID starting with a5eb2fe712743b6df5b92327728adda9631d0629b363beac574ec4f9a9265142 not found: ID does not exist" containerID="a5eb2fe712743b6df5b92327728adda9631d0629b363beac574ec4f9a9265142" Sep 30 09:30:40 crc kubenswrapper[4810]: I0930 09:30:40.099099 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5eb2fe712743b6df5b92327728adda9631d0629b363beac574ec4f9a9265142"} err="failed to get container status \"a5eb2fe712743b6df5b92327728adda9631d0629b363beac574ec4f9a9265142\": rpc error: code = NotFound desc = could not find container \"a5eb2fe712743b6df5b92327728adda9631d0629b363beac574ec4f9a9265142\": container with ID starting with a5eb2fe712743b6df5b92327728adda9631d0629b363beac574ec4f9a9265142 not found: ID does not exist" Sep 30 09:30:40 crc kubenswrapper[4810]: I0930 09:30:40.099123 4810 scope.go:117] "RemoveContainer" containerID="922c6de228a2e65b09a8aab043431ebcdf7e1b38834b20279a284ae965cda435" Sep 30 09:30:40 crc kubenswrapper[4810]: E0930 09:30:40.099378 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"922c6de228a2e65b09a8aab043431ebcdf7e1b38834b20279a284ae965cda435\": container with ID starting with 922c6de228a2e65b09a8aab043431ebcdf7e1b38834b20279a284ae965cda435 not found: ID does not exist" containerID="922c6de228a2e65b09a8aab043431ebcdf7e1b38834b20279a284ae965cda435" Sep 30 09:30:40 crc kubenswrapper[4810]: I0930 09:30:40.099402 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"922c6de228a2e65b09a8aab043431ebcdf7e1b38834b20279a284ae965cda435"} err="failed to get container status \"922c6de228a2e65b09a8aab043431ebcdf7e1b38834b20279a284ae965cda435\": rpc error: code = NotFound desc = could not find container \"922c6de228a2e65b09a8aab043431ebcdf7e1b38834b20279a284ae965cda435\": container with ID starting with 922c6de228a2e65b09a8aab043431ebcdf7e1b38834b20279a284ae965cda435 not found: ID does not exist" Sep 30 09:30:40 crc kubenswrapper[4810]: I0930 09:30:40.173037 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fxchk"] Sep 30 09:30:40 crc kubenswrapper[4810]: I0930 09:30:40.182084 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fxchk"] Sep 30 09:30:41 crc kubenswrapper[4810]: I0930 09:30:41.345765 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e49d38-ec21-403d-9f29-97a3b9415cab" path="/var/lib/kubelet/pods/72e49d38-ec21-403d-9f29-97a3b9415cab/volumes" Sep 30 09:30:45 crc kubenswrapper[4810]: I0930 09:30:45.912117 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:30:45 crc kubenswrapper[4810]: I0930 09:30:45.912804 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:31:15 crc kubenswrapper[4810]: I0930 09:31:15.911940 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:31:15 crc kubenswrapper[4810]: I0930 09:31:15.912516 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:31:15 crc kubenswrapper[4810]: I0930 09:31:15.912571 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 09:31:15 crc kubenswrapper[4810]: I0930 09:31:15.913468 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 09:31:15 crc kubenswrapper[4810]: I0930 09:31:15.913550 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" gracePeriod=600 Sep 30 09:31:16 crc kubenswrapper[4810]: E0930 09:31:16.049252 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:31:16 crc kubenswrapper[4810]: I0930 09:31:16.898419 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" exitCode=0 Sep 30 09:31:16 crc kubenswrapper[4810]: I0930 09:31:16.898485 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68"} Sep 30 09:31:16 crc kubenswrapper[4810]: I0930 09:31:16.898537 4810 scope.go:117] "RemoveContainer" containerID="fdefe6392dc2bfa1f689117fddb7ef6ba2f43e8dcedf0924a7ce3f58a2613787" Sep 30 09:31:16 crc kubenswrapper[4810]: I0930 09:31:16.899493 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:31:16 crc kubenswrapper[4810]: E0930 09:31:16.899926 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:31:32 crc kubenswrapper[4810]: I0930 09:31:32.307516 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:31:32 crc kubenswrapper[4810]: E0930 09:31:32.309745 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:31:44 crc kubenswrapper[4810]: I0930 09:31:44.307277 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:31:44 crc kubenswrapper[4810]: E0930 09:31:44.308075 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:31:56 crc kubenswrapper[4810]: I0930 09:31:56.306573 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:31:56 crc kubenswrapper[4810]: E0930 09:31:56.307285 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:32:11 crc kubenswrapper[4810]: I0930 09:32:11.306406 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:32:11 crc kubenswrapper[4810]: E0930 09:32:11.307356 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:32:23 crc kubenswrapper[4810]: I0930 09:32:23.308581 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:32:23 crc kubenswrapper[4810]: E0930 09:32:23.310799 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:32:34 crc kubenswrapper[4810]: I0930 09:32:34.307132 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:32:34 crc kubenswrapper[4810]: E0930 09:32:34.308227 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:32:46 crc kubenswrapper[4810]: I0930 09:32:46.306915 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:32:46 crc kubenswrapper[4810]: E0930 09:32:46.307840 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:33:00 crc kubenswrapper[4810]: I0930 09:33:00.306532 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:33:00 crc kubenswrapper[4810]: E0930 09:33:00.307232 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:33:14 crc kubenswrapper[4810]: I0930 09:33:14.307171 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:33:14 crc kubenswrapper[4810]: E0930 09:33:14.308320 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.199797 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n27mq"] Sep 30 09:33:19 crc kubenswrapper[4810]: E0930 09:33:19.200883 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e49d38-ec21-403d-9f29-97a3b9415cab" containerName="extract-content" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.200901 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e49d38-ec21-403d-9f29-97a3b9415cab" containerName="extract-content" Sep 30 09:33:19 crc kubenswrapper[4810]: E0930 09:33:19.200918 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e49d38-ec21-403d-9f29-97a3b9415cab" containerName="extract-utilities" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.200926 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e49d38-ec21-403d-9f29-97a3b9415cab" containerName="extract-utilities" Sep 30 09:33:19 crc kubenswrapper[4810]: E0930 09:33:19.200939 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e49d38-ec21-403d-9f29-97a3b9415cab" containerName="registry-server" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.200949 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e49d38-ec21-403d-9f29-97a3b9415cab" containerName="registry-server" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.201252 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e49d38-ec21-403d-9f29-97a3b9415cab" containerName="registry-server" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.203136 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.212843 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n27mq"] Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.320058 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtkvq\" (UniqueName: \"kubernetes.io/projected/294d9ae0-8482-4091-80c2-ac77506868dd-kube-api-access-vtkvq\") pod \"community-operators-n27mq\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.320218 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-catalog-content\") pod \"community-operators-n27mq\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.320259 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-utilities\") pod \"community-operators-n27mq\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.423057 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtkvq\" (UniqueName: \"kubernetes.io/projected/294d9ae0-8482-4091-80c2-ac77506868dd-kube-api-access-vtkvq\") pod \"community-operators-n27mq\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.423389 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-catalog-content\") pod \"community-operators-n27mq\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.423455 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-utilities\") pod \"community-operators-n27mq\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.424155 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-catalog-content\") pod \"community-operators-n27mq\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.424460 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-utilities\") pod \"community-operators-n27mq\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.452919 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtkvq\" (UniqueName: \"kubernetes.io/projected/294d9ae0-8482-4091-80c2-ac77506868dd-kube-api-access-vtkvq\") pod \"community-operators-n27mq\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:19 crc kubenswrapper[4810]: I0930 09:33:19.536607 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:20 crc kubenswrapper[4810]: I0930 09:33:20.084802 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n27mq"] Sep 30 09:33:20 crc kubenswrapper[4810]: I0930 09:33:20.213526 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n27mq" event={"ID":"294d9ae0-8482-4091-80c2-ac77506868dd","Type":"ContainerStarted","Data":"4483998b7b6525ef9b747875e47ff3c0fa2778a4a41912c57ac897f8ce9aa3da"} Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.226004 4810 generic.go:334] "Generic (PLEG): container finished" podID="294d9ae0-8482-4091-80c2-ac77506868dd" containerID="7f64aa297822f14d696a5cfe4fc69fa93ef9b9e03a38b63da4f05fbb30ccfe3a" exitCode=0 Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.226112 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n27mq" event={"ID":"294d9ae0-8482-4091-80c2-ac77506868dd","Type":"ContainerDied","Data":"7f64aa297822f14d696a5cfe4fc69fa93ef9b9e03a38b63da4f05fbb30ccfe3a"} Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.414140 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x7zhw"] Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.421886 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.469802 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7zhw"] Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.571359 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b679059-88b3-4a0f-8623-cb5a2a2c0e82-utilities\") pod \"redhat-operators-x7zhw\" (UID: \"6b679059-88b3-4a0f-8623-cb5a2a2c0e82\") " pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.571440 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b679059-88b3-4a0f-8623-cb5a2a2c0e82-catalog-content\") pod \"redhat-operators-x7zhw\" (UID: \"6b679059-88b3-4a0f-8623-cb5a2a2c0e82\") " pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.571537 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r49k6\" (UniqueName: \"kubernetes.io/projected/6b679059-88b3-4a0f-8623-cb5a2a2c0e82-kube-api-access-r49k6\") pod \"redhat-operators-x7zhw\" (UID: \"6b679059-88b3-4a0f-8623-cb5a2a2c0e82\") " pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.673487 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r49k6\" (UniqueName: \"kubernetes.io/projected/6b679059-88b3-4a0f-8623-cb5a2a2c0e82-kube-api-access-r49k6\") pod \"redhat-operators-x7zhw\" (UID: \"6b679059-88b3-4a0f-8623-cb5a2a2c0e82\") " pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.673647 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b679059-88b3-4a0f-8623-cb5a2a2c0e82-utilities\") pod \"redhat-operators-x7zhw\" (UID: \"6b679059-88b3-4a0f-8623-cb5a2a2c0e82\") " pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.673703 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b679059-88b3-4a0f-8623-cb5a2a2c0e82-catalog-content\") pod \"redhat-operators-x7zhw\" (UID: \"6b679059-88b3-4a0f-8623-cb5a2a2c0e82\") " pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.674157 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b679059-88b3-4a0f-8623-cb5a2a2c0e82-catalog-content\") pod \"redhat-operators-x7zhw\" (UID: \"6b679059-88b3-4a0f-8623-cb5a2a2c0e82\") " pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.674348 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b679059-88b3-4a0f-8623-cb5a2a2c0e82-utilities\") pod \"redhat-operators-x7zhw\" (UID: \"6b679059-88b3-4a0f-8623-cb5a2a2c0e82\") " pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.694832 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r49k6\" (UniqueName: \"kubernetes.io/projected/6b679059-88b3-4a0f-8623-cb5a2a2c0e82-kube-api-access-r49k6\") pod \"redhat-operators-x7zhw\" (UID: \"6b679059-88b3-4a0f-8623-cb5a2a2c0e82\") " pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:21 crc kubenswrapper[4810]: I0930 09:33:21.748194 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:22 crc kubenswrapper[4810]: I0930 09:33:22.231977 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7zhw"] Sep 30 09:33:23 crc kubenswrapper[4810]: I0930 09:33:23.246687 4810 generic.go:334] "Generic (PLEG): container finished" podID="6b679059-88b3-4a0f-8623-cb5a2a2c0e82" containerID="6473662d81ed2df753174f54787d930042f2a7e82ac5f0d25a9abac30f560aa2" exitCode=0 Sep 30 09:33:23 crc kubenswrapper[4810]: I0930 09:33:23.246823 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7zhw" event={"ID":"6b679059-88b3-4a0f-8623-cb5a2a2c0e82","Type":"ContainerDied","Data":"6473662d81ed2df753174f54787d930042f2a7e82ac5f0d25a9abac30f560aa2"} Sep 30 09:33:23 crc kubenswrapper[4810]: I0930 09:33:23.246857 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7zhw" event={"ID":"6b679059-88b3-4a0f-8623-cb5a2a2c0e82","Type":"ContainerStarted","Data":"61e573f9014a7b00f23f0f73d06ac61892c691e191b58261005f708c22f920c9"} Sep 30 09:33:23 crc kubenswrapper[4810]: I0930 09:33:23.255088 4810 generic.go:334] "Generic (PLEG): container finished" podID="294d9ae0-8482-4091-80c2-ac77506868dd" containerID="631b8923ba4bfa436706e0f30791b9a0c31bc34639b923f10b824dfcb30d1add" exitCode=0 Sep 30 09:33:23 crc kubenswrapper[4810]: I0930 09:33:23.255174 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n27mq" event={"ID":"294d9ae0-8482-4091-80c2-ac77506868dd","Type":"ContainerDied","Data":"631b8923ba4bfa436706e0f30791b9a0c31bc34639b923f10b824dfcb30d1add"} Sep 30 09:33:24 crc kubenswrapper[4810]: I0930 09:33:24.281403 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n27mq" event={"ID":"294d9ae0-8482-4091-80c2-ac77506868dd","Type":"ContainerStarted","Data":"3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02"} Sep 30 09:33:24 crc kubenswrapper[4810]: I0930 09:33:24.334518 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n27mq" podStartSLOduration=2.82971373 podStartE2EDuration="5.334494111s" podCreationTimestamp="2025-09-30 09:33:19 +0000 UTC" firstStartedPulling="2025-09-30 09:33:21.229767379 +0000 UTC m=+5424.681966646" lastFinishedPulling="2025-09-30 09:33:23.73454774 +0000 UTC m=+5427.186747027" observedRunningTime="2025-09-30 09:33:24.324160542 +0000 UTC m=+5427.776359819" watchObservedRunningTime="2025-09-30 09:33:24.334494111 +0000 UTC m=+5427.786693378" Sep 30 09:33:28 crc kubenswrapper[4810]: I0930 09:33:28.306432 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:33:28 crc kubenswrapper[4810]: E0930 09:33:28.307197 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:33:29 crc kubenswrapper[4810]: I0930 09:33:29.537469 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:29 crc kubenswrapper[4810]: I0930 09:33:29.537823 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:29 crc kubenswrapper[4810]: I0930 09:33:29.586964 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:30 crc kubenswrapper[4810]: I0930 09:33:30.414832 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:30 crc kubenswrapper[4810]: I0930 09:33:30.488846 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n27mq"] Sep 30 09:33:32 crc kubenswrapper[4810]: I0930 09:33:32.365448 4810 generic.go:334] "Generic (PLEG): container finished" podID="6b679059-88b3-4a0f-8623-cb5a2a2c0e82" containerID="04078757801ed28a65fcbfe814c8a00588170ec74508a33b8afe9d4efe541209" exitCode=0 Sep 30 09:33:32 crc kubenswrapper[4810]: I0930 09:33:32.365525 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7zhw" event={"ID":"6b679059-88b3-4a0f-8623-cb5a2a2c0e82","Type":"ContainerDied","Data":"04078757801ed28a65fcbfe814c8a00588170ec74508a33b8afe9d4efe541209"} Sep 30 09:33:32 crc kubenswrapper[4810]: I0930 09:33:32.366057 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n27mq" podUID="294d9ae0-8482-4091-80c2-ac77506868dd" containerName="registry-server" containerID="cri-o://3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02" gracePeriod=2 Sep 30 09:33:32 crc kubenswrapper[4810]: I0930 09:33:32.905682 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.064673 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtkvq\" (UniqueName: \"kubernetes.io/projected/294d9ae0-8482-4091-80c2-ac77506868dd-kube-api-access-vtkvq\") pod \"294d9ae0-8482-4091-80c2-ac77506868dd\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.064807 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-catalog-content\") pod \"294d9ae0-8482-4091-80c2-ac77506868dd\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.064830 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-utilities\") pod \"294d9ae0-8482-4091-80c2-ac77506868dd\" (UID: \"294d9ae0-8482-4091-80c2-ac77506868dd\") " Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.065417 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-utilities" (OuterVolumeSpecName: "utilities") pod "294d9ae0-8482-4091-80c2-ac77506868dd" (UID: "294d9ae0-8482-4091-80c2-ac77506868dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.066389 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.070120 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/294d9ae0-8482-4091-80c2-ac77506868dd-kube-api-access-vtkvq" (OuterVolumeSpecName: "kube-api-access-vtkvq") pod "294d9ae0-8482-4091-80c2-ac77506868dd" (UID: "294d9ae0-8482-4091-80c2-ac77506868dd"). InnerVolumeSpecName "kube-api-access-vtkvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.106481 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "294d9ae0-8482-4091-80c2-ac77506868dd" (UID: "294d9ae0-8482-4091-80c2-ac77506868dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.168144 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtkvq\" (UniqueName: \"kubernetes.io/projected/294d9ae0-8482-4091-80c2-ac77506868dd-kube-api-access-vtkvq\") on node \"crc\" DevicePath \"\"" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.168186 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294d9ae0-8482-4091-80c2-ac77506868dd-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.377112 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7zhw" event={"ID":"6b679059-88b3-4a0f-8623-cb5a2a2c0e82","Type":"ContainerStarted","Data":"d9dd4c543285243fd4a2780af0445873516d2f2196e47f9b859fd8bd11d60ca3"} Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.379704 4810 generic.go:334] "Generic (PLEG): container finished" podID="294d9ae0-8482-4091-80c2-ac77506868dd" containerID="3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02" exitCode=0 Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.379797 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n27mq" event={"ID":"294d9ae0-8482-4091-80c2-ac77506868dd","Type":"ContainerDied","Data":"3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02"} Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.379831 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n27mq" event={"ID":"294d9ae0-8482-4091-80c2-ac77506868dd","Type":"ContainerDied","Data":"4483998b7b6525ef9b747875e47ff3c0fa2778a4a41912c57ac897f8ce9aa3da"} Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.379860 4810 scope.go:117] "RemoveContainer" containerID="3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.379927 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n27mq" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.398203 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x7zhw" podStartSLOduration=2.686922712 podStartE2EDuration="12.398187962s" podCreationTimestamp="2025-09-30 09:33:21 +0000 UTC" firstStartedPulling="2025-09-30 09:33:23.250097361 +0000 UTC m=+5426.702296628" lastFinishedPulling="2025-09-30 09:33:32.961362601 +0000 UTC m=+5436.413561878" observedRunningTime="2025-09-30 09:33:33.396005479 +0000 UTC m=+5436.848204756" watchObservedRunningTime="2025-09-30 09:33:33.398187962 +0000 UTC m=+5436.850387229" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.403599 4810 scope.go:117] "RemoveContainer" containerID="631b8923ba4bfa436706e0f30791b9a0c31bc34639b923f10b824dfcb30d1add" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.431471 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n27mq"] Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.431540 4810 scope.go:117] "RemoveContainer" containerID="7f64aa297822f14d696a5cfe4fc69fa93ef9b9e03a38b63da4f05fbb30ccfe3a" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.441276 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n27mq"] Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.454754 4810 scope.go:117] "RemoveContainer" containerID="3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02" Sep 30 09:33:33 crc kubenswrapper[4810]: E0930 09:33:33.455256 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02\": container with ID starting with 3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02 not found: ID does not exist" containerID="3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.455423 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02"} err="failed to get container status \"3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02\": rpc error: code = NotFound desc = could not find container \"3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02\": container with ID starting with 3aecff8f79256b2b1d3d81d0a2bde9875b925754fcd63afbd40bdea56a911d02 not found: ID does not exist" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.455651 4810 scope.go:117] "RemoveContainer" containerID="631b8923ba4bfa436706e0f30791b9a0c31bc34639b923f10b824dfcb30d1add" Sep 30 09:33:33 crc kubenswrapper[4810]: E0930 09:33:33.456189 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"631b8923ba4bfa436706e0f30791b9a0c31bc34639b923f10b824dfcb30d1add\": container with ID starting with 631b8923ba4bfa436706e0f30791b9a0c31bc34639b923f10b824dfcb30d1add not found: ID does not exist" containerID="631b8923ba4bfa436706e0f30791b9a0c31bc34639b923f10b824dfcb30d1add" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.456303 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"631b8923ba4bfa436706e0f30791b9a0c31bc34639b923f10b824dfcb30d1add"} err="failed to get container status \"631b8923ba4bfa436706e0f30791b9a0c31bc34639b923f10b824dfcb30d1add\": rpc error: code = NotFound desc = could not find container \"631b8923ba4bfa436706e0f30791b9a0c31bc34639b923f10b824dfcb30d1add\": container with ID starting with 631b8923ba4bfa436706e0f30791b9a0c31bc34639b923f10b824dfcb30d1add not found: ID does not exist" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.456413 4810 scope.go:117] "RemoveContainer" containerID="7f64aa297822f14d696a5cfe4fc69fa93ef9b9e03a38b63da4f05fbb30ccfe3a" Sep 30 09:33:33 crc kubenswrapper[4810]: E0930 09:33:33.456830 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f64aa297822f14d696a5cfe4fc69fa93ef9b9e03a38b63da4f05fbb30ccfe3a\": container with ID starting with 7f64aa297822f14d696a5cfe4fc69fa93ef9b9e03a38b63da4f05fbb30ccfe3a not found: ID does not exist" containerID="7f64aa297822f14d696a5cfe4fc69fa93ef9b9e03a38b63da4f05fbb30ccfe3a" Sep 30 09:33:33 crc kubenswrapper[4810]: I0930 09:33:33.456864 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f64aa297822f14d696a5cfe4fc69fa93ef9b9e03a38b63da4f05fbb30ccfe3a"} err="failed to get container status \"7f64aa297822f14d696a5cfe4fc69fa93ef9b9e03a38b63da4f05fbb30ccfe3a\": rpc error: code = NotFound desc = could not find container \"7f64aa297822f14d696a5cfe4fc69fa93ef9b9e03a38b63da4f05fbb30ccfe3a\": container with ID starting with 7f64aa297822f14d696a5cfe4fc69fa93ef9b9e03a38b63da4f05fbb30ccfe3a not found: ID does not exist" Sep 30 09:33:33 crc kubenswrapper[4810]: E0930 09:33:33.518695 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod294d9ae0_8482_4091_80c2_ac77506868dd.slice/crio-4483998b7b6525ef9b747875e47ff3c0fa2778a4a41912c57ac897f8ce9aa3da\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod294d9ae0_8482_4091_80c2_ac77506868dd.slice\": RecentStats: unable to find data in memory cache]" Sep 30 09:33:35 crc kubenswrapper[4810]: I0930 09:33:35.319720 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="294d9ae0-8482-4091-80c2-ac77506868dd" path="/var/lib/kubelet/pods/294d9ae0-8482-4091-80c2-ac77506868dd/volumes" Sep 30 09:33:40 crc kubenswrapper[4810]: I0930 09:33:40.307034 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:33:40 crc kubenswrapper[4810]: E0930 09:33:40.308127 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:33:41 crc kubenswrapper[4810]: I0930 09:33:41.749107 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:41 crc kubenswrapper[4810]: I0930 09:33:41.749437 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:41 crc kubenswrapper[4810]: I0930 09:33:41.821945 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:42 crc kubenswrapper[4810]: I0930 09:33:42.533184 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x7zhw" Sep 30 09:33:42 crc kubenswrapper[4810]: I0930 09:33:42.612783 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7zhw"] Sep 30 09:33:42 crc kubenswrapper[4810]: I0930 09:33:42.664724 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k6m7g"] Sep 30 09:33:42 crc kubenswrapper[4810]: I0930 09:33:42.665244 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k6m7g" podUID="fa6eca49-c738-41e6-bfda-134db27352bc" containerName="registry-server" containerID="cri-o://e8b89276c8b32befe99256a44f1054ef92c71828bc922ddb2fa52bcaf99fb577" gracePeriod=2 Sep 30 09:33:43 crc kubenswrapper[4810]: I0930 09:33:43.497188 4810 generic.go:334] "Generic (PLEG): container finished" podID="fa6eca49-c738-41e6-bfda-134db27352bc" containerID="e8b89276c8b32befe99256a44f1054ef92c71828bc922ddb2fa52bcaf99fb577" exitCode=0 Sep 30 09:33:43 crc kubenswrapper[4810]: I0930 09:33:43.497297 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6m7g" event={"ID":"fa6eca49-c738-41e6-bfda-134db27352bc","Type":"ContainerDied","Data":"e8b89276c8b32befe99256a44f1054ef92c71828bc922ddb2fa52bcaf99fb577"} Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.309967 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.418327 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-catalog-content\") pod \"fa6eca49-c738-41e6-bfda-134db27352bc\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.418551 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xvh2\" (UniqueName: \"kubernetes.io/projected/fa6eca49-c738-41e6-bfda-134db27352bc-kube-api-access-4xvh2\") pod \"fa6eca49-c738-41e6-bfda-134db27352bc\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.418708 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-utilities\") pod \"fa6eca49-c738-41e6-bfda-134db27352bc\" (UID: \"fa6eca49-c738-41e6-bfda-134db27352bc\") " Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.419141 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-utilities" (OuterVolumeSpecName: "utilities") pod "fa6eca49-c738-41e6-bfda-134db27352bc" (UID: "fa6eca49-c738-41e6-bfda-134db27352bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.419320 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.423869 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa6eca49-c738-41e6-bfda-134db27352bc-kube-api-access-4xvh2" (OuterVolumeSpecName: "kube-api-access-4xvh2") pod "fa6eca49-c738-41e6-bfda-134db27352bc" (UID: "fa6eca49-c738-41e6-bfda-134db27352bc"). InnerVolumeSpecName "kube-api-access-4xvh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.495604 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa6eca49-c738-41e6-bfda-134db27352bc" (UID: "fa6eca49-c738-41e6-bfda-134db27352bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.509498 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6m7g" event={"ID":"fa6eca49-c738-41e6-bfda-134db27352bc","Type":"ContainerDied","Data":"2657acd00503727a1239796c2616b378995bbcf6d84605c85221172dd45ce18a"} Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.509553 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6m7g" Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.509573 4810 scope.go:117] "RemoveContainer" containerID="e8b89276c8b32befe99256a44f1054ef92c71828bc922ddb2fa52bcaf99fb577" Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.521115 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa6eca49-c738-41e6-bfda-134db27352bc-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.521140 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xvh2\" (UniqueName: \"kubernetes.io/projected/fa6eca49-c738-41e6-bfda-134db27352bc-kube-api-access-4xvh2\") on node \"crc\" DevicePath \"\"" Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.531728 4810 scope.go:117] "RemoveContainer" containerID="52ef7609e74159d0646144324670786db8077f570001a01270879c9aed1b2b3b" Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.551867 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k6m7g"] Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.565967 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k6m7g"] Sep 30 09:33:44 crc kubenswrapper[4810]: I0930 09:33:44.568711 4810 scope.go:117] "RemoveContainer" containerID="964f43b6b2db911d9d9c9e00446f1072f098b352ac9e90a281d16401f1e94703" Sep 30 09:33:45 crc kubenswrapper[4810]: I0930 09:33:45.320027 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa6eca49-c738-41e6-bfda-134db27352bc" path="/var/lib/kubelet/pods/fa6eca49-c738-41e6-bfda-134db27352bc/volumes" Sep 30 09:33:52 crc kubenswrapper[4810]: I0930 09:33:52.307536 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:33:52 crc kubenswrapper[4810]: E0930 09:33:52.308598 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:34:06 crc kubenswrapper[4810]: I0930 09:34:06.308756 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:34:06 crc kubenswrapper[4810]: E0930 09:34:06.309691 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:34:19 crc kubenswrapper[4810]: I0930 09:34:19.307889 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:34:19 crc kubenswrapper[4810]: E0930 09:34:19.309237 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:34:33 crc kubenswrapper[4810]: I0930 09:34:33.306978 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:34:33 crc kubenswrapper[4810]: E0930 09:34:33.307811 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:34:45 crc kubenswrapper[4810]: I0930 09:34:45.153862 4810 generic.go:334] "Generic (PLEG): container finished" podID="78242ee7-a6aa-4087-832e-4834a7f4751f" containerID="70d98fa1bc91e767b2be403c958203e9f54e59dd68eb8aefaa7cb90743a0f8ae" exitCode=0 Sep 30 09:34:45 crc kubenswrapper[4810]: I0930 09:34:45.153980 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"78242ee7-a6aa-4087-832e-4834a7f4751f","Type":"ContainerDied","Data":"70d98fa1bc91e767b2be403c958203e9f54e59dd68eb8aefaa7cb90743a0f8ae"} Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.567655 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.624898 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ca-certs\") pod \"78242ee7-a6aa-4087-832e-4834a7f4751f\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.624970 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ssh-key\") pod \"78242ee7-a6aa-4087-832e-4834a7f4751f\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.625066 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-config-data\") pod \"78242ee7-a6aa-4087-832e-4834a7f4751f\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.625170 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-temporary\") pod \"78242ee7-a6aa-4087-832e-4834a7f4751f\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.625190 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-workdir\") pod \"78242ee7-a6aa-4087-832e-4834a7f4751f\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.625216 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config-secret\") pod \"78242ee7-a6aa-4087-832e-4834a7f4751f\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.625237 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl7hm\" (UniqueName: \"kubernetes.io/projected/78242ee7-a6aa-4087-832e-4834a7f4751f-kube-api-access-bl7hm\") pod \"78242ee7-a6aa-4087-832e-4834a7f4751f\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.625267 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"78242ee7-a6aa-4087-832e-4834a7f4751f\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.625332 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config\") pod \"78242ee7-a6aa-4087-832e-4834a7f4751f\" (UID: \"78242ee7-a6aa-4087-832e-4834a7f4751f\") " Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.626672 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "78242ee7-a6aa-4087-832e-4834a7f4751f" (UID: "78242ee7-a6aa-4087-832e-4834a7f4751f"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.627028 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-config-data" (OuterVolumeSpecName: "config-data") pod "78242ee7-a6aa-4087-832e-4834a7f4751f" (UID: "78242ee7-a6aa-4087-832e-4834a7f4751f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.630973 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "78242ee7-a6aa-4087-832e-4834a7f4751f" (UID: "78242ee7-a6aa-4087-832e-4834a7f4751f"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.631688 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78242ee7-a6aa-4087-832e-4834a7f4751f-kube-api-access-bl7hm" (OuterVolumeSpecName: "kube-api-access-bl7hm") pod "78242ee7-a6aa-4087-832e-4834a7f4751f" (UID: "78242ee7-a6aa-4087-832e-4834a7f4751f"). InnerVolumeSpecName "kube-api-access-bl7hm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.646130 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "78242ee7-a6aa-4087-832e-4834a7f4751f" (UID: "78242ee7-a6aa-4087-832e-4834a7f4751f"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.661964 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "78242ee7-a6aa-4087-832e-4834a7f4751f" (UID: "78242ee7-a6aa-4087-832e-4834a7f4751f"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.663456 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "78242ee7-a6aa-4087-832e-4834a7f4751f" (UID: "78242ee7-a6aa-4087-832e-4834a7f4751f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.676204 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "78242ee7-a6aa-4087-832e-4834a7f4751f" (UID: "78242ee7-a6aa-4087-832e-4834a7f4751f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.679946 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "78242ee7-a6aa-4087-832e-4834a7f4751f" (UID: "78242ee7-a6aa-4087-832e-4834a7f4751f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.727340 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.727384 4810 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.727400 4810 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/78242ee7-a6aa-4087-832e-4834a7f4751f-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.727412 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.727421 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl7hm\" (UniqueName: \"kubernetes.io/projected/78242ee7-a6aa-4087-832e-4834a7f4751f-kube-api-access-bl7hm\") on node \"crc\" DevicePath \"\"" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.727453 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.727462 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/78242ee7-a6aa-4087-832e-4834a7f4751f-openstack-config\") on node \"crc\" DevicePath \"\"" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.727470 4810 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ca-certs\") on node \"crc\" DevicePath \"\"" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.727478 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78242ee7-a6aa-4087-832e-4834a7f4751f-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.753057 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Sep 30 09:34:46 crc kubenswrapper[4810]: I0930 09:34:46.830105 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Sep 30 09:34:47 crc kubenswrapper[4810]: I0930 09:34:47.182230 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"78242ee7-a6aa-4087-832e-4834a7f4751f","Type":"ContainerDied","Data":"e30963a5616e7727f82bd8527dbc09b62c422e4faae0b2eaf8d32e5655813bf8"} Sep 30 09:34:47 crc kubenswrapper[4810]: I0930 09:34:47.182283 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e30963a5616e7727f82bd8527dbc09b62c422e4faae0b2eaf8d32e5655813bf8" Sep 30 09:34:47 crc kubenswrapper[4810]: I0930 09:34:47.182397 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 09:34:48 crc kubenswrapper[4810]: I0930 09:34:48.306804 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:34:48 crc kubenswrapper[4810]: E0930 09:34:48.307676 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.184855 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 30 09:34:58 crc kubenswrapper[4810]: E0930 09:34:58.186028 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa6eca49-c738-41e6-bfda-134db27352bc" containerName="extract-utilities" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.186052 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa6eca49-c738-41e6-bfda-134db27352bc" containerName="extract-utilities" Sep 30 09:34:58 crc kubenswrapper[4810]: E0930 09:34:58.186096 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294d9ae0-8482-4091-80c2-ac77506868dd" containerName="extract-utilities" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.186106 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="294d9ae0-8482-4091-80c2-ac77506868dd" containerName="extract-utilities" Sep 30 09:34:58 crc kubenswrapper[4810]: E0930 09:34:58.186129 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294d9ae0-8482-4091-80c2-ac77506868dd" containerName="extract-content" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.186141 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="294d9ae0-8482-4091-80c2-ac77506868dd" containerName="extract-content" Sep 30 09:34:58 crc kubenswrapper[4810]: E0930 09:34:58.186158 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa6eca49-c738-41e6-bfda-134db27352bc" containerName="extract-content" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.186171 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa6eca49-c738-41e6-bfda-134db27352bc" containerName="extract-content" Sep 30 09:34:58 crc kubenswrapper[4810]: E0930 09:34:58.186191 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa6eca49-c738-41e6-bfda-134db27352bc" containerName="registry-server" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.186201 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa6eca49-c738-41e6-bfda-134db27352bc" containerName="registry-server" Sep 30 09:34:58 crc kubenswrapper[4810]: E0930 09:34:58.186223 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78242ee7-a6aa-4087-832e-4834a7f4751f" containerName="tempest-tests-tempest-tests-runner" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.186233 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="78242ee7-a6aa-4087-832e-4834a7f4751f" containerName="tempest-tests-tempest-tests-runner" Sep 30 09:34:58 crc kubenswrapper[4810]: E0930 09:34:58.186254 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294d9ae0-8482-4091-80c2-ac77506868dd" containerName="registry-server" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.186285 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="294d9ae0-8482-4091-80c2-ac77506868dd" containerName="registry-server" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.186628 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="294d9ae0-8482-4091-80c2-ac77506868dd" containerName="registry-server" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.186658 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="78242ee7-a6aa-4087-832e-4834a7f4751f" containerName="tempest-tests-tempest-tests-runner" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.186679 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa6eca49-c738-41e6-bfda-134db27352bc" containerName="registry-server" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.187717 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.190018 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vms5j" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.200180 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.288142 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg982\" (UniqueName: \"kubernetes.io/projected/1d54c982-ce8e-43cb-ab4f-69a3cbaebdda-kube-api-access-gg982\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1d54c982-ce8e-43cb-ab4f-69a3cbaebdda\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.288223 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1d54c982-ce8e-43cb-ab4f-69a3cbaebdda\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.390557 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg982\" (UniqueName: \"kubernetes.io/projected/1d54c982-ce8e-43cb-ab4f-69a3cbaebdda-kube-api-access-gg982\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1d54c982-ce8e-43cb-ab4f-69a3cbaebdda\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.390650 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1d54c982-ce8e-43cb-ab4f-69a3cbaebdda\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.394138 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1d54c982-ce8e-43cb-ab4f-69a3cbaebdda\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.424668 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg982\" (UniqueName: \"kubernetes.io/projected/1d54c982-ce8e-43cb-ab4f-69a3cbaebdda-kube-api-access-gg982\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1d54c982-ce8e-43cb-ab4f-69a3cbaebdda\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.437246 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1d54c982-ce8e-43cb-ab4f-69a3cbaebdda\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 09:34:58 crc kubenswrapper[4810]: I0930 09:34:58.544180 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 09:34:59 crc kubenswrapper[4810]: I0930 09:34:59.005853 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 30 09:34:59 crc kubenswrapper[4810]: I0930 09:34:59.008733 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 09:34:59 crc kubenswrapper[4810]: I0930 09:34:59.318245 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"1d54c982-ce8e-43cb-ab4f-69a3cbaebdda","Type":"ContainerStarted","Data":"9de0f8df08717484d03bf1058c4e8f41000fe3893cd192a08556749ccc804077"} Sep 30 09:35:00 crc kubenswrapper[4810]: I0930 09:35:00.331205 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"1d54c982-ce8e-43cb-ab4f-69a3cbaebdda","Type":"ContainerStarted","Data":"a5e525a4c72b56718ed1b03d37dc109be71fe55c42e4d127729887f77a15e434"} Sep 30 09:35:00 crc kubenswrapper[4810]: I0930 09:35:00.350368 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.358792964 podStartE2EDuration="2.350346767s" podCreationTimestamp="2025-09-30 09:34:58 +0000 UTC" firstStartedPulling="2025-09-30 09:34:59.008439085 +0000 UTC m=+5522.460638362" lastFinishedPulling="2025-09-30 09:34:59.999992878 +0000 UTC m=+5523.452192165" observedRunningTime="2025-09-30 09:35:00.348700579 +0000 UTC m=+5523.800899846" watchObservedRunningTime="2025-09-30 09:35:00.350346767 +0000 UTC m=+5523.802546034" Sep 30 09:35:02 crc kubenswrapper[4810]: I0930 09:35:02.305999 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:35:02 crc kubenswrapper[4810]: E0930 09:35:02.306555 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:35:15 crc kubenswrapper[4810]: I0930 09:35:15.307375 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:35:15 crc kubenswrapper[4810]: E0930 09:35:15.308654 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.408060 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-48zlz/must-gather-jj5v6"] Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.410769 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/must-gather-jj5v6" Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.414747 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-48zlz"/"default-dockercfg-wbf59" Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.414782 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-48zlz"/"openshift-service-ca.crt" Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.415062 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-48zlz"/"kube-root-ca.crt" Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.423850 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-48zlz/must-gather-jj5v6"] Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.562681 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/439949d3-6a38-4e79-bded-bee33d6884ef-must-gather-output\") pod \"must-gather-jj5v6\" (UID: \"439949d3-6a38-4e79-bded-bee33d6884ef\") " pod="openshift-must-gather-48zlz/must-gather-jj5v6" Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.563081 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-427g2\" (UniqueName: \"kubernetes.io/projected/439949d3-6a38-4e79-bded-bee33d6884ef-kube-api-access-427g2\") pod \"must-gather-jj5v6\" (UID: \"439949d3-6a38-4e79-bded-bee33d6884ef\") " pod="openshift-must-gather-48zlz/must-gather-jj5v6" Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.665316 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/439949d3-6a38-4e79-bded-bee33d6884ef-must-gather-output\") pod \"must-gather-jj5v6\" (UID: \"439949d3-6a38-4e79-bded-bee33d6884ef\") " pod="openshift-must-gather-48zlz/must-gather-jj5v6" Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.665464 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-427g2\" (UniqueName: \"kubernetes.io/projected/439949d3-6a38-4e79-bded-bee33d6884ef-kube-api-access-427g2\") pod \"must-gather-jj5v6\" (UID: \"439949d3-6a38-4e79-bded-bee33d6884ef\") " pod="openshift-must-gather-48zlz/must-gather-jj5v6" Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.666194 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/439949d3-6a38-4e79-bded-bee33d6884ef-must-gather-output\") pod \"must-gather-jj5v6\" (UID: \"439949d3-6a38-4e79-bded-bee33d6884ef\") " pod="openshift-must-gather-48zlz/must-gather-jj5v6" Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.690038 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-427g2\" (UniqueName: \"kubernetes.io/projected/439949d3-6a38-4e79-bded-bee33d6884ef-kube-api-access-427g2\") pod \"must-gather-jj5v6\" (UID: \"439949d3-6a38-4e79-bded-bee33d6884ef\") " pod="openshift-must-gather-48zlz/must-gather-jj5v6" Sep 30 09:35:19 crc kubenswrapper[4810]: I0930 09:35:19.734357 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/must-gather-jj5v6" Sep 30 09:35:20 crc kubenswrapper[4810]: I0930 09:35:20.217547 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-48zlz/must-gather-jj5v6"] Sep 30 09:35:20 crc kubenswrapper[4810]: I0930 09:35:20.530028 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/must-gather-jj5v6" event={"ID":"439949d3-6a38-4e79-bded-bee33d6884ef","Type":"ContainerStarted","Data":"423a7ce095d39eb32b7956a33c1d3ff724eb6e3d6184f6c02c67aa775f01918c"} Sep 30 09:35:29 crc kubenswrapper[4810]: I0930 09:35:29.643170 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/must-gather-jj5v6" event={"ID":"439949d3-6a38-4e79-bded-bee33d6884ef","Type":"ContainerStarted","Data":"300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5"} Sep 30 09:35:29 crc kubenswrapper[4810]: I0930 09:35:29.643946 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/must-gather-jj5v6" event={"ID":"439949d3-6a38-4e79-bded-bee33d6884ef","Type":"ContainerStarted","Data":"d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f"} Sep 30 09:35:29 crc kubenswrapper[4810]: I0930 09:35:29.664702 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-48zlz/must-gather-jj5v6" podStartSLOduration=2.182254842 podStartE2EDuration="10.664681722s" podCreationTimestamp="2025-09-30 09:35:19 +0000 UTC" firstStartedPulling="2025-09-30 09:35:20.219970885 +0000 UTC m=+5543.672170152" lastFinishedPulling="2025-09-30 09:35:28.702397765 +0000 UTC m=+5552.154597032" observedRunningTime="2025-09-30 09:35:29.656825884 +0000 UTC m=+5553.109025151" watchObservedRunningTime="2025-09-30 09:35:29.664681722 +0000 UTC m=+5553.116880989" Sep 30 09:35:30 crc kubenswrapper[4810]: I0930 09:35:30.306111 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:35:30 crc kubenswrapper[4810]: E0930 09:35:30.306379 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:35:33 crc kubenswrapper[4810]: I0930 09:35:33.185582 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-48zlz/crc-debug-mc5kz"] Sep 30 09:35:33 crc kubenswrapper[4810]: I0930 09:35:33.187672 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-mc5kz" Sep 30 09:35:33 crc kubenswrapper[4810]: I0930 09:35:33.267371 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcnmn\" (UniqueName: \"kubernetes.io/projected/d64b2c6f-545e-4937-90bf-922fb4d60fb7-kube-api-access-mcnmn\") pod \"crc-debug-mc5kz\" (UID: \"d64b2c6f-545e-4937-90bf-922fb4d60fb7\") " pod="openshift-must-gather-48zlz/crc-debug-mc5kz" Sep 30 09:35:33 crc kubenswrapper[4810]: I0930 09:35:33.267621 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d64b2c6f-545e-4937-90bf-922fb4d60fb7-host\") pod \"crc-debug-mc5kz\" (UID: \"d64b2c6f-545e-4937-90bf-922fb4d60fb7\") " pod="openshift-must-gather-48zlz/crc-debug-mc5kz" Sep 30 09:35:33 crc kubenswrapper[4810]: I0930 09:35:33.370190 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcnmn\" (UniqueName: \"kubernetes.io/projected/d64b2c6f-545e-4937-90bf-922fb4d60fb7-kube-api-access-mcnmn\") pod \"crc-debug-mc5kz\" (UID: \"d64b2c6f-545e-4937-90bf-922fb4d60fb7\") " pod="openshift-must-gather-48zlz/crc-debug-mc5kz" Sep 30 09:35:33 crc kubenswrapper[4810]: I0930 09:35:33.370858 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d64b2c6f-545e-4937-90bf-922fb4d60fb7-host\") pod \"crc-debug-mc5kz\" (UID: \"d64b2c6f-545e-4937-90bf-922fb4d60fb7\") " pod="openshift-must-gather-48zlz/crc-debug-mc5kz" Sep 30 09:35:33 crc kubenswrapper[4810]: I0930 09:35:33.370938 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d64b2c6f-545e-4937-90bf-922fb4d60fb7-host\") pod \"crc-debug-mc5kz\" (UID: \"d64b2c6f-545e-4937-90bf-922fb4d60fb7\") " pod="openshift-must-gather-48zlz/crc-debug-mc5kz" Sep 30 09:35:33 crc kubenswrapper[4810]: I0930 09:35:33.388162 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcnmn\" (UniqueName: \"kubernetes.io/projected/d64b2c6f-545e-4937-90bf-922fb4d60fb7-kube-api-access-mcnmn\") pod \"crc-debug-mc5kz\" (UID: \"d64b2c6f-545e-4937-90bf-922fb4d60fb7\") " pod="openshift-must-gather-48zlz/crc-debug-mc5kz" Sep 30 09:35:33 crc kubenswrapper[4810]: I0930 09:35:33.508404 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-mc5kz" Sep 30 09:35:33 crc kubenswrapper[4810]: W0930 09:35:33.558939 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd64b2c6f_545e_4937_90bf_922fb4d60fb7.slice/crio-c6eb146c8088444000d6373b10e77c240ca2bae36e90129d4be70fb0c8a99e54 WatchSource:0}: Error finding container c6eb146c8088444000d6373b10e77c240ca2bae36e90129d4be70fb0c8a99e54: Status 404 returned error can't find the container with id c6eb146c8088444000d6373b10e77c240ca2bae36e90129d4be70fb0c8a99e54 Sep 30 09:35:33 crc kubenswrapper[4810]: I0930 09:35:33.688306 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/crc-debug-mc5kz" event={"ID":"d64b2c6f-545e-4937-90bf-922fb4d60fb7","Type":"ContainerStarted","Data":"c6eb146c8088444000d6373b10e77c240ca2bae36e90129d4be70fb0c8a99e54"} Sep 30 09:35:42 crc kubenswrapper[4810]: I0930 09:35:42.307307 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:35:42 crc kubenswrapper[4810]: E0930 09:35:42.308219 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:35:44 crc kubenswrapper[4810]: I0930 09:35:44.815794 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/crc-debug-mc5kz" event={"ID":"d64b2c6f-545e-4937-90bf-922fb4d60fb7","Type":"ContainerStarted","Data":"c0d4a7c966617e668b0012d053edc1135ec5cc9832f877c0275a1a66d47860b8"} Sep 30 09:35:44 crc kubenswrapper[4810]: I0930 09:35:44.836923 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-48zlz/crc-debug-mc5kz" podStartSLOduration=1.362315918 podStartE2EDuration="11.836903336s" podCreationTimestamp="2025-09-30 09:35:33 +0000 UTC" firstStartedPulling="2025-09-30 09:35:33.56132685 +0000 UTC m=+5557.013526117" lastFinishedPulling="2025-09-30 09:35:44.035914268 +0000 UTC m=+5567.488113535" observedRunningTime="2025-09-30 09:35:44.829192002 +0000 UTC m=+5568.281391269" watchObservedRunningTime="2025-09-30 09:35:44.836903336 +0000 UTC m=+5568.289102603" Sep 30 09:35:53 crc kubenswrapper[4810]: I0930 09:35:53.308362 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:35:53 crc kubenswrapper[4810]: E0930 09:35:53.308842 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:36:08 crc kubenswrapper[4810]: I0930 09:36:08.306316 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:36:08 crc kubenswrapper[4810]: E0930 09:36:08.307081 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:36:22 crc kubenswrapper[4810]: I0930 09:36:22.306191 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:36:23 crc kubenswrapper[4810]: I0930 09:36:23.194780 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"3cfe1e0b8905c291915720129de0e23bdd06ea6549e1c8ca139a56a2878a94dd"} Sep 30 09:36:59 crc kubenswrapper[4810]: I0930 09:36:59.528116 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c6d89f69d-k9g44_23454188-6622-4e7c-90f0-2fddf031738e/barbican-api/0.log" Sep 30 09:36:59 crc kubenswrapper[4810]: I0930 09:36:59.553947 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c6d89f69d-k9g44_23454188-6622-4e7c-90f0-2fddf031738e/barbican-api-log/0.log" Sep 30 09:36:59 crc kubenswrapper[4810]: I0930 09:36:59.723456 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5dff5594b-jj7gs_d2c5a1aa-c611-4702-a0bb-27693ebb7c7a/barbican-keystone-listener/0.log" Sep 30 09:36:59 crc kubenswrapper[4810]: I0930 09:36:59.819784 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5dff5594b-jj7gs_d2c5a1aa-c611-4702-a0bb-27693ebb7c7a/barbican-keystone-listener-log/0.log" Sep 30 09:36:59 crc kubenswrapper[4810]: I0930 09:36:59.920044 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-687444b9dc-hxqlq_ff274965-178c-4eff-9841-494634e711f0/barbican-worker/0.log" Sep 30 09:37:00 crc kubenswrapper[4810]: I0930 09:37:00.010144 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-687444b9dc-hxqlq_ff274965-178c-4eff-9841-494634e711f0/barbican-worker-log/0.log" Sep 30 09:37:00 crc kubenswrapper[4810]: I0930 09:37:00.188017 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz_6465dd71-b2e8-4a11-90a8-406ddd763a0c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:00 crc kubenswrapper[4810]: I0930 09:37:00.396978 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ae076b6e-09cd-4911-89c3-e6edb61516ae/ceilometer-central-agent/0.log" Sep 30 09:37:00 crc kubenswrapper[4810]: I0930 09:37:00.470086 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ae076b6e-09cd-4911-89c3-e6edb61516ae/ceilometer-notification-agent/0.log" Sep 30 09:37:00 crc kubenswrapper[4810]: I0930 09:37:00.516625 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ae076b6e-09cd-4911-89c3-e6edb61516ae/proxy-httpd/0.log" Sep 30 09:37:00 crc kubenswrapper[4810]: I0930 09:37:00.607634 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ae076b6e-09cd-4911-89c3-e6edb61516ae/sg-core/0.log" Sep 30 09:37:00 crc kubenswrapper[4810]: I0930 09:37:00.833134 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_58537813-eea7-4a58-a900-f3473797d7d2/cinder-api-log/0.log" Sep 30 09:37:00 crc kubenswrapper[4810]: I0930 09:37:00.898308 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_58537813-eea7-4a58-a900-f3473797d7d2/cinder-api/0.log" Sep 30 09:37:01 crc kubenswrapper[4810]: I0930 09:37:01.015084 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e10f507-33b5-4477-882f-6c0b3271034d/cinder-scheduler/0.log" Sep 30 09:37:01 crc kubenswrapper[4810]: I0930 09:37:01.131686 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e10f507-33b5-4477-882f-6c0b3271034d/probe/0.log" Sep 30 09:37:01 crc kubenswrapper[4810]: I0930 09:37:01.274861 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-22dr4_459ff4e4-7043-4e6e-9774-97dab472459c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:01 crc kubenswrapper[4810]: I0930 09:37:01.355283 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx_80653750-cd3d-4e95-b4a1-4909c325c34c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:01 crc kubenswrapper[4810]: I0930 09:37:01.563073 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-585f78d44c-nlbbz_cf951386-0eef-4a7f-9247-b854b9a39d7b/init/0.log" Sep 30 09:37:01 crc kubenswrapper[4810]: I0930 09:37:01.739157 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-585f78d44c-nlbbz_cf951386-0eef-4a7f-9247-b854b9a39d7b/init/0.log" Sep 30 09:37:01 crc kubenswrapper[4810]: I0930 09:37:01.922719 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-585f78d44c-nlbbz_cf951386-0eef-4a7f-9247-b854b9a39d7b/dnsmasq-dns/0.log" Sep 30 09:37:01 crc kubenswrapper[4810]: I0930 09:37:01.977368 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq_3f13a98d-9385-4e17-a8ef-fd0dec1866ac/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:02 crc kubenswrapper[4810]: I0930 09:37:02.127884 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f35de952-dcbd-422d-bbdb-b8995de342a1/glance-httpd/0.log" Sep 30 09:37:02 crc kubenswrapper[4810]: I0930 09:37:02.160230 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f35de952-dcbd-422d-bbdb-b8995de342a1/glance-log/0.log" Sep 30 09:37:02 crc kubenswrapper[4810]: I0930 09:37:02.323069 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_77634c10-6e69-4f6c-9090-be8a59588745/glance-httpd/0.log" Sep 30 09:37:02 crc kubenswrapper[4810]: I0930 09:37:02.365569 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_77634c10-6e69-4f6c-9090-be8a59588745/glance-log/0.log" Sep 30 09:37:02 crc kubenswrapper[4810]: I0930 09:37:02.724698 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d769bd5f4-zkxj4_a71102a5-388b-4dcb-bd80-db92ed28f1f5/horizon/0.log" Sep 30 09:37:02 crc kubenswrapper[4810]: I0930 09:37:02.790802 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-2rltf_d2e22e53-6826-4a58-abb0-eb2d279d2b1c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:02 crc kubenswrapper[4810]: I0930 09:37:02.994634 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-9gnn7_75b17470-760f-4fa6-9550-c2281ed5ae6b/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:03 crc kubenswrapper[4810]: I0930 09:37:03.109940 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29320381-wgv24_dd37005c-311f-4aef-9f91-bc01d5d62249/keystone-cron/0.log" Sep 30 09:37:03 crc kubenswrapper[4810]: I0930 09:37:03.258300 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d769bd5f4-zkxj4_a71102a5-388b-4dcb-bd80-db92ed28f1f5/horizon-log/0.log" Sep 30 09:37:03 crc kubenswrapper[4810]: I0930 09:37:03.467407 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b0778b82-d141-47f3-b766-a3df63f09d12/kube-state-metrics/0.log" Sep 30 09:37:03 crc kubenswrapper[4810]: I0930 09:37:03.585253 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd_9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:03 crc kubenswrapper[4810]: I0930 09:37:03.796296 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-f95cc94f6-tz5ks_2d05e909-d2bc-49d7-af7b-0736de4c3f95/keystone-api/0.log" Sep 30 09:37:04 crc kubenswrapper[4810]: I0930 09:37:04.279047 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f769c46d7-mgh2q_34fe741b-5b18-4fe8-9f54-1fd3e62d9e24/neutron-httpd/0.log" Sep 30 09:37:04 crc kubenswrapper[4810]: I0930 09:37:04.315572 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z_5de6b629-1489-4099-89bf-7773b0bdaeb8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:04 crc kubenswrapper[4810]: I0930 09:37:04.321786 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f769c46d7-mgh2q_34fe741b-5b18-4fe8-9f54-1fd3e62d9e24/neutron-api/0.log" Sep 30 09:37:05 crc kubenswrapper[4810]: I0930 09:37:05.235652 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_8f67f47e-4e54-44b8-9596-c612858accdd/nova-cell0-conductor-conductor/0.log" Sep 30 09:37:05 crc kubenswrapper[4810]: I0930 09:37:05.771345 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_4342c2f0-9d27-4211-a302-a1822255f011/nova-cell1-conductor-conductor/0.log" Sep 30 09:37:06 crc kubenswrapper[4810]: I0930 09:37:06.179447 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_7e2cfe82-5249-4207-9178-f4ebb56a9f93/nova-api-log/0.log" Sep 30 09:37:06 crc kubenswrapper[4810]: I0930 09:37:06.324832 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_7e2cfe82-5249-4207-9178-f4ebb56a9f93/nova-api-api/0.log" Sep 30 09:37:06 crc kubenswrapper[4810]: I0930 09:37:06.361128 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_fc321aeb-240d-414e-91de-ba95671d6556/nova-cell1-novncproxy-novncproxy/0.log" Sep 30 09:37:06 crc kubenswrapper[4810]: I0930 09:37:06.637526 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-v6q68_8a25c107-d531-4a79-97e9-7201d6a8485a/nova-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:06 crc kubenswrapper[4810]: I0930 09:37:06.764653 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_3d733771-3677-4a2e-9eda-86144efccf8d/nova-metadata-log/0.log" Sep 30 09:37:07 crc kubenswrapper[4810]: I0930 09:37:07.287034 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_97a5417c-b645-430c-ada6-ae56ca4aa813/nova-scheduler-scheduler/0.log" Sep 30 09:37:07 crc kubenswrapper[4810]: I0930 09:37:07.339842 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cb4aa97-f850-4e7f-a1e1-046bd5235109/mysql-bootstrap/0.log" Sep 30 09:37:07 crc kubenswrapper[4810]: I0930 09:37:07.534867 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cb4aa97-f850-4e7f-a1e1-046bd5235109/galera/0.log" Sep 30 09:37:07 crc kubenswrapper[4810]: I0930 09:37:07.567475 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cb4aa97-f850-4e7f-a1e1-046bd5235109/mysql-bootstrap/0.log" Sep 30 09:37:07 crc kubenswrapper[4810]: I0930 09:37:07.750966 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3ca490c3-ff2f-4553-9e66-4fb456d3a1f2/mysql-bootstrap/0.log" Sep 30 09:37:07 crc kubenswrapper[4810]: I0930 09:37:07.973320 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3ca490c3-ff2f-4553-9e66-4fb456d3a1f2/mysql-bootstrap/0.log" Sep 30 09:37:08 crc kubenswrapper[4810]: I0930 09:37:08.063885 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3ca490c3-ff2f-4553-9e66-4fb456d3a1f2/galera/0.log" Sep 30 09:37:08 crc kubenswrapper[4810]: I0930 09:37:08.268809 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f0aa786d-3e79-423b-82d1-5e9c70ca7c72/openstackclient/0.log" Sep 30 09:37:08 crc kubenswrapper[4810]: I0930 09:37:08.501594 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-9q2d8_af34271b-1a44-4240-88ee-b86490744353/ovn-controller/0.log" Sep 30 09:37:08 crc kubenswrapper[4810]: I0930 09:37:08.689562 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-6wrgl_4b9a3337-b2c2-49c8-9755-81660bb506af/openstack-network-exporter/0.log" Sep 30 09:37:08 crc kubenswrapper[4810]: I0930 09:37:08.945934 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mt698_384b3d69-0ca1-4729-86f7-6d4455fb7ccc/ovsdb-server-init/0.log" Sep 30 09:37:08 crc kubenswrapper[4810]: I0930 09:37:08.975366 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_3d733771-3677-4a2e-9eda-86144efccf8d/nova-metadata-metadata/0.log" Sep 30 09:37:09 crc kubenswrapper[4810]: I0930 09:37:09.131586 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mt698_384b3d69-0ca1-4729-86f7-6d4455fb7ccc/ovsdb-server-init/0.log" Sep 30 09:37:09 crc kubenswrapper[4810]: I0930 09:37:09.211116 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mt698_384b3d69-0ca1-4729-86f7-6d4455fb7ccc/ovsdb-server/0.log" Sep 30 09:37:09 crc kubenswrapper[4810]: I0930 09:37:09.426238 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-5xm2r_dbee734f-4300-41cc-9bb9-fa005daa1df0/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:09 crc kubenswrapper[4810]: I0930 09:37:09.561506 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mt698_384b3d69-0ca1-4729-86f7-6d4455fb7ccc/ovs-vswitchd/0.log" Sep 30 09:37:09 crc kubenswrapper[4810]: I0930 09:37:09.630416 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_306e4bf0-015f-4c7a-ba49-efda86e72442/openstack-network-exporter/0.log" Sep 30 09:37:09 crc kubenswrapper[4810]: I0930 09:37:09.738767 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_306e4bf0-015f-4c7a-ba49-efda86e72442/ovn-northd/0.log" Sep 30 09:37:09 crc kubenswrapper[4810]: I0930 09:37:09.898840 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c94c8fec-2762-469d-86f9-a4aca5c242e5/openstack-network-exporter/0.log" Sep 30 09:37:09 crc kubenswrapper[4810]: I0930 09:37:09.995328 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c94c8fec-2762-469d-86f9-a4aca5c242e5/ovsdbserver-nb/0.log" Sep 30 09:37:10 crc kubenswrapper[4810]: I0930 09:37:10.124778 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ba947eb7-f01f-40ef-b276-b5b73fee9e0a/openstack-network-exporter/0.log" Sep 30 09:37:10 crc kubenswrapper[4810]: I0930 09:37:10.210378 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ba947eb7-f01f-40ef-b276-b5b73fee9e0a/ovsdbserver-sb/0.log" Sep 30 09:37:10 crc kubenswrapper[4810]: I0930 09:37:10.532417 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-54b65d888b-2hkpr_35128b6f-4e9c-414b-847f-773c1418f5e2/placement-api/0.log" Sep 30 09:37:10 crc kubenswrapper[4810]: I0930 09:37:10.745332 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-54b65d888b-2hkpr_35128b6f-4e9c-414b-847f-773c1418f5e2/placement-log/0.log" Sep 30 09:37:10 crc kubenswrapper[4810]: I0930 09:37:10.771529 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ad43140a-2b1c-443b-9595-901c71e14f0d/init-config-reloader/0.log" Sep 30 09:37:10 crc kubenswrapper[4810]: I0930 09:37:10.927821 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ad43140a-2b1c-443b-9595-901c71e14f0d/config-reloader/0.log" Sep 30 09:37:10 crc kubenswrapper[4810]: I0930 09:37:10.972651 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ad43140a-2b1c-443b-9595-901c71e14f0d/init-config-reloader/0.log" Sep 30 09:37:11 crc kubenswrapper[4810]: I0930 09:37:11.000122 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ad43140a-2b1c-443b-9595-901c71e14f0d/prometheus/0.log" Sep 30 09:37:11 crc kubenswrapper[4810]: I0930 09:37:11.113670 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ad43140a-2b1c-443b-9595-901c71e14f0d/thanos-sidecar/0.log" Sep 30 09:37:11 crc kubenswrapper[4810]: I0930 09:37:11.266487 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0344203-4b62-409c-af24-0e619dfad39a/setup-container/0.log" Sep 30 09:37:11 crc kubenswrapper[4810]: I0930 09:37:11.425868 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0344203-4b62-409c-af24-0e619dfad39a/setup-container/0.log" Sep 30 09:37:11 crc kubenswrapper[4810]: I0930 09:37:11.460754 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0344203-4b62-409c-af24-0e619dfad39a/rabbitmq/0.log" Sep 30 09:37:11 crc kubenswrapper[4810]: I0930 09:37:11.629345 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_fcf1eb6a-de6d-48ac-bbf5-4e015c452e18/setup-container/0.log" Sep 30 09:37:11 crc kubenswrapper[4810]: I0930 09:37:11.825668 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_fcf1eb6a-de6d-48ac-bbf5-4e015c452e18/setup-container/0.log" Sep 30 09:37:11 crc kubenswrapper[4810]: I0930 09:37:11.872218 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_fcf1eb6a-de6d-48ac-bbf5-4e015c452e18/rabbitmq/0.log" Sep 30 09:37:12 crc kubenswrapper[4810]: I0930 09:37:12.022597 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b695439c-7bde-408e-b90a-0aa0ff0bc494/setup-container/0.log" Sep 30 09:37:12 crc kubenswrapper[4810]: I0930 09:37:12.200348 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b695439c-7bde-408e-b90a-0aa0ff0bc494/rabbitmq/0.log" Sep 30 09:37:12 crc kubenswrapper[4810]: I0930 09:37:12.260982 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b695439c-7bde-408e-b90a-0aa0ff0bc494/setup-container/0.log" Sep 30 09:37:12 crc kubenswrapper[4810]: I0930 09:37:12.493091 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp_6da1c758-a003-49dd-95de-f47d13e11c0a/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:12 crc kubenswrapper[4810]: I0930 09:37:12.512479 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-vhxfz_dc05f1d1-59b0-45d8-8a52-e0cc456a1207/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:12 crc kubenswrapper[4810]: I0930 09:37:12.689630 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r_93d196dc-dc56-4116-bb59-d76601947afd/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:12 crc kubenswrapper[4810]: I0930 09:37:12.953099 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-v8zdn_48233a79-011d-4b63-8f0c-885c651ddbee/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:13 crc kubenswrapper[4810]: I0930 09:37:13.019028 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-6gw9x_d5b6b2ff-6366-48ba-9328-7d99a9da7e08/ssh-known-hosts-edpm-deployment/0.log" Sep 30 09:37:13 crc kubenswrapper[4810]: I0930 09:37:13.298712 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5cd556568c-9c6km_dc1aa89c-399a-4db2-a0b1-df1185b14c48/proxy-server/0.log" Sep 30 09:37:13 crc kubenswrapper[4810]: I0930 09:37:13.515871 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-zr8dp_cf76de28-c7b8-474f-93a7-5af536e817e8/swift-ring-rebalance/0.log" Sep 30 09:37:13 crc kubenswrapper[4810]: I0930 09:37:13.518940 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5cd556568c-9c6km_dc1aa89c-399a-4db2-a0b1-df1185b14c48/proxy-httpd/0.log" Sep 30 09:37:13 crc kubenswrapper[4810]: I0930 09:37:13.716721 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/account-auditor/0.log" Sep 30 09:37:13 crc kubenswrapper[4810]: I0930 09:37:13.725065 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/account-reaper/0.log" Sep 30 09:37:13 crc kubenswrapper[4810]: I0930 09:37:13.922452 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/account-replicator/0.log" Sep 30 09:37:13 crc kubenswrapper[4810]: I0930 09:37:13.971861 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/account-server/0.log" Sep 30 09:37:14 crc kubenswrapper[4810]: I0930 09:37:14.011971 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/container-auditor/0.log" Sep 30 09:37:14 crc kubenswrapper[4810]: I0930 09:37:14.164419 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/container-replicator/0.log" Sep 30 09:37:14 crc kubenswrapper[4810]: I0930 09:37:14.183824 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/container-server/0.log" Sep 30 09:37:14 crc kubenswrapper[4810]: I0930 09:37:14.228962 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/container-updater/0.log" Sep 30 09:37:14 crc kubenswrapper[4810]: I0930 09:37:14.397934 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/object-expirer/0.log" Sep 30 09:37:14 crc kubenswrapper[4810]: I0930 09:37:14.398829 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/object-auditor/0.log" Sep 30 09:37:14 crc kubenswrapper[4810]: I0930 09:37:14.439548 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/object-replicator/0.log" Sep 30 09:37:14 crc kubenswrapper[4810]: I0930 09:37:14.575528 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/object-server/0.log" Sep 30 09:37:14 crc kubenswrapper[4810]: I0930 09:37:14.644065 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/object-updater/0.log" Sep 30 09:37:14 crc kubenswrapper[4810]: I0930 09:37:14.742854 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/rsync/0.log" Sep 30 09:37:14 crc kubenswrapper[4810]: I0930 09:37:14.777684 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/swift-recon-cron/0.log" Sep 30 09:37:15 crc kubenswrapper[4810]: I0930 09:37:15.016693 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r_ba3931a9-05a1-47f8-a7f7-f41a477a164b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:15 crc kubenswrapper[4810]: I0930 09:37:15.191425 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_78242ee7-a6aa-4087-832e-4834a7f4751f/tempest-tests-tempest-tests-runner/0.log" Sep 30 09:37:15 crc kubenswrapper[4810]: I0930 09:37:15.364205 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_1d54c982-ce8e-43cb-ab4f-69a3cbaebdda/test-operator-logs-container/0.log" Sep 30 09:37:15 crc kubenswrapper[4810]: I0930 09:37:15.660742 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b_a86849c2-c174-4b69-84a0-281bb1afda6d/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:37:16 crc kubenswrapper[4810]: I0930 09:37:16.800200 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_d04899be-cfe2-4870-ad3c-735148e7bcdd/watcher-applier/0.log" Sep 30 09:37:16 crc kubenswrapper[4810]: I0930 09:37:16.994895 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_32ede37e-d419-4c76-932e-af1de102a9ad/watcher-api-log/0.log" Sep 30 09:37:20 crc kubenswrapper[4810]: I0930 09:37:20.677968 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_5c9ec55c-ddec-4cab-8041-6b5a9c192212/watcher-decision-engine/0.log" Sep 30 09:37:21 crc kubenswrapper[4810]: I0930 09:37:21.204936 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_32ede37e-d419-4c76-932e-af1de102a9ad/watcher-api/0.log" Sep 30 09:37:30 crc kubenswrapper[4810]: I0930 09:37:30.625990 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_f947e702-a475-413a-8e34-08702649bdae/memcached/0.log" Sep 30 09:37:59 crc kubenswrapper[4810]: I0930 09:37:59.318938 4810 generic.go:334] "Generic (PLEG): container finished" podID="d64b2c6f-545e-4937-90bf-922fb4d60fb7" containerID="c0d4a7c966617e668b0012d053edc1135ec5cc9832f877c0275a1a66d47860b8" exitCode=0 Sep 30 09:37:59 crc kubenswrapper[4810]: I0930 09:37:59.332795 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/crc-debug-mc5kz" event={"ID":"d64b2c6f-545e-4937-90bf-922fb4d60fb7","Type":"ContainerDied","Data":"c0d4a7c966617e668b0012d053edc1135ec5cc9832f877c0275a1a66d47860b8"} Sep 30 09:38:00 crc kubenswrapper[4810]: I0930 09:38:00.465230 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-mc5kz" Sep 30 09:38:00 crc kubenswrapper[4810]: I0930 09:38:00.505217 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-48zlz/crc-debug-mc5kz"] Sep 30 09:38:00 crc kubenswrapper[4810]: I0930 09:38:00.512674 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-48zlz/crc-debug-mc5kz"] Sep 30 09:38:00 crc kubenswrapper[4810]: I0930 09:38:00.635329 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcnmn\" (UniqueName: \"kubernetes.io/projected/d64b2c6f-545e-4937-90bf-922fb4d60fb7-kube-api-access-mcnmn\") pod \"d64b2c6f-545e-4937-90bf-922fb4d60fb7\" (UID: \"d64b2c6f-545e-4937-90bf-922fb4d60fb7\") " Sep 30 09:38:00 crc kubenswrapper[4810]: I0930 09:38:00.635493 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d64b2c6f-545e-4937-90bf-922fb4d60fb7-host\") pod \"d64b2c6f-545e-4937-90bf-922fb4d60fb7\" (UID: \"d64b2c6f-545e-4937-90bf-922fb4d60fb7\") " Sep 30 09:38:00 crc kubenswrapper[4810]: I0930 09:38:00.635769 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d64b2c6f-545e-4937-90bf-922fb4d60fb7-host" (OuterVolumeSpecName: "host") pod "d64b2c6f-545e-4937-90bf-922fb4d60fb7" (UID: "d64b2c6f-545e-4937-90bf-922fb4d60fb7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 09:38:00 crc kubenswrapper[4810]: I0930 09:38:00.636704 4810 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d64b2c6f-545e-4937-90bf-922fb4d60fb7-host\") on node \"crc\" DevicePath \"\"" Sep 30 09:38:00 crc kubenswrapper[4810]: I0930 09:38:00.646607 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64b2c6f-545e-4937-90bf-922fb4d60fb7-kube-api-access-mcnmn" (OuterVolumeSpecName: "kube-api-access-mcnmn") pod "d64b2c6f-545e-4937-90bf-922fb4d60fb7" (UID: "d64b2c6f-545e-4937-90bf-922fb4d60fb7"). InnerVolumeSpecName "kube-api-access-mcnmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:38:00 crc kubenswrapper[4810]: I0930 09:38:00.737995 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcnmn\" (UniqueName: \"kubernetes.io/projected/d64b2c6f-545e-4937-90bf-922fb4d60fb7-kube-api-access-mcnmn\") on node \"crc\" DevicePath \"\"" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.321838 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d64b2c6f-545e-4937-90bf-922fb4d60fb7" path="/var/lib/kubelet/pods/d64b2c6f-545e-4937-90bf-922fb4d60fb7/volumes" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.344640 4810 scope.go:117] "RemoveContainer" containerID="c0d4a7c966617e668b0012d053edc1135ec5cc9832f877c0275a1a66d47860b8" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.344718 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-mc5kz" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.711517 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-48zlz/crc-debug-92q5f"] Sep 30 09:38:01 crc kubenswrapper[4810]: E0930 09:38:01.711918 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64b2c6f-545e-4937-90bf-922fb4d60fb7" containerName="container-00" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.711929 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64b2c6f-545e-4937-90bf-922fb4d60fb7" containerName="container-00" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.712118 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64b2c6f-545e-4937-90bf-922fb4d60fb7" containerName="container-00" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.712866 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-92q5f" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.859702 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e93a8f1-6809-4d54-8231-cc351fe8d592-host\") pod \"crc-debug-92q5f\" (UID: \"2e93a8f1-6809-4d54-8231-cc351fe8d592\") " pod="openshift-must-gather-48zlz/crc-debug-92q5f" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.860183 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmwq8\" (UniqueName: \"kubernetes.io/projected/2e93a8f1-6809-4d54-8231-cc351fe8d592-kube-api-access-kmwq8\") pod \"crc-debug-92q5f\" (UID: \"2e93a8f1-6809-4d54-8231-cc351fe8d592\") " pod="openshift-must-gather-48zlz/crc-debug-92q5f" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.962062 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmwq8\" (UniqueName: \"kubernetes.io/projected/2e93a8f1-6809-4d54-8231-cc351fe8d592-kube-api-access-kmwq8\") pod \"crc-debug-92q5f\" (UID: \"2e93a8f1-6809-4d54-8231-cc351fe8d592\") " pod="openshift-must-gather-48zlz/crc-debug-92q5f" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.962186 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e93a8f1-6809-4d54-8231-cc351fe8d592-host\") pod \"crc-debug-92q5f\" (UID: \"2e93a8f1-6809-4d54-8231-cc351fe8d592\") " pod="openshift-must-gather-48zlz/crc-debug-92q5f" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.962335 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e93a8f1-6809-4d54-8231-cc351fe8d592-host\") pod \"crc-debug-92q5f\" (UID: \"2e93a8f1-6809-4d54-8231-cc351fe8d592\") " pod="openshift-must-gather-48zlz/crc-debug-92q5f" Sep 30 09:38:01 crc kubenswrapper[4810]: I0930 09:38:01.985524 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmwq8\" (UniqueName: \"kubernetes.io/projected/2e93a8f1-6809-4d54-8231-cc351fe8d592-kube-api-access-kmwq8\") pod \"crc-debug-92q5f\" (UID: \"2e93a8f1-6809-4d54-8231-cc351fe8d592\") " pod="openshift-must-gather-48zlz/crc-debug-92q5f" Sep 30 09:38:02 crc kubenswrapper[4810]: I0930 09:38:02.029706 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-92q5f" Sep 30 09:38:02 crc kubenswrapper[4810]: W0930 09:38:02.065480 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e93a8f1_6809_4d54_8231_cc351fe8d592.slice/crio-65f6c9b960ec6ffb97573d1d7bdcc5acf3759a2e35f3d6e2eb9cd625de9f24af WatchSource:0}: Error finding container 65f6c9b960ec6ffb97573d1d7bdcc5acf3759a2e35f3d6e2eb9cd625de9f24af: Status 404 returned error can't find the container with id 65f6c9b960ec6ffb97573d1d7bdcc5acf3759a2e35f3d6e2eb9cd625de9f24af Sep 30 09:38:02 crc kubenswrapper[4810]: I0930 09:38:02.361442 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/crc-debug-92q5f" event={"ID":"2e93a8f1-6809-4d54-8231-cc351fe8d592","Type":"ContainerStarted","Data":"88a0559833e2aadd014203d1bdfe41327bbf9ad4a049400611aa1a291ed42f96"} Sep 30 09:38:02 crc kubenswrapper[4810]: I0930 09:38:02.361780 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/crc-debug-92q5f" event={"ID":"2e93a8f1-6809-4d54-8231-cc351fe8d592","Type":"ContainerStarted","Data":"65f6c9b960ec6ffb97573d1d7bdcc5acf3759a2e35f3d6e2eb9cd625de9f24af"} Sep 30 09:38:02 crc kubenswrapper[4810]: I0930 09:38:02.378951 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-48zlz/crc-debug-92q5f" podStartSLOduration=1.378929398 podStartE2EDuration="1.378929398s" podCreationTimestamp="2025-09-30 09:38:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 09:38:02.375170299 +0000 UTC m=+5705.827369566" watchObservedRunningTime="2025-09-30 09:38:02.378929398 +0000 UTC m=+5705.831128665" Sep 30 09:38:03 crc kubenswrapper[4810]: I0930 09:38:03.376321 4810 generic.go:334] "Generic (PLEG): container finished" podID="2e93a8f1-6809-4d54-8231-cc351fe8d592" containerID="88a0559833e2aadd014203d1bdfe41327bbf9ad4a049400611aa1a291ed42f96" exitCode=0 Sep 30 09:38:03 crc kubenswrapper[4810]: I0930 09:38:03.376365 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/crc-debug-92q5f" event={"ID":"2e93a8f1-6809-4d54-8231-cc351fe8d592","Type":"ContainerDied","Data":"88a0559833e2aadd014203d1bdfe41327bbf9ad4a049400611aa1a291ed42f96"} Sep 30 09:38:04 crc kubenswrapper[4810]: I0930 09:38:04.497650 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-92q5f" Sep 30 09:38:04 crc kubenswrapper[4810]: I0930 09:38:04.602253 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e93a8f1-6809-4d54-8231-cc351fe8d592-host\") pod \"2e93a8f1-6809-4d54-8231-cc351fe8d592\" (UID: \"2e93a8f1-6809-4d54-8231-cc351fe8d592\") " Sep 30 09:38:04 crc kubenswrapper[4810]: I0930 09:38:04.602320 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmwq8\" (UniqueName: \"kubernetes.io/projected/2e93a8f1-6809-4d54-8231-cc351fe8d592-kube-api-access-kmwq8\") pod \"2e93a8f1-6809-4d54-8231-cc351fe8d592\" (UID: \"2e93a8f1-6809-4d54-8231-cc351fe8d592\") " Sep 30 09:38:04 crc kubenswrapper[4810]: I0930 09:38:04.602329 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e93a8f1-6809-4d54-8231-cc351fe8d592-host" (OuterVolumeSpecName: "host") pod "2e93a8f1-6809-4d54-8231-cc351fe8d592" (UID: "2e93a8f1-6809-4d54-8231-cc351fe8d592"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 09:38:04 crc kubenswrapper[4810]: I0930 09:38:04.602699 4810 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e93a8f1-6809-4d54-8231-cc351fe8d592-host\") on node \"crc\" DevicePath \"\"" Sep 30 09:38:04 crc kubenswrapper[4810]: I0930 09:38:04.608761 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e93a8f1-6809-4d54-8231-cc351fe8d592-kube-api-access-kmwq8" (OuterVolumeSpecName: "kube-api-access-kmwq8") pod "2e93a8f1-6809-4d54-8231-cc351fe8d592" (UID: "2e93a8f1-6809-4d54-8231-cc351fe8d592"). InnerVolumeSpecName "kube-api-access-kmwq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:38:04 crc kubenswrapper[4810]: I0930 09:38:04.703891 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmwq8\" (UniqueName: \"kubernetes.io/projected/2e93a8f1-6809-4d54-8231-cc351fe8d592-kube-api-access-kmwq8\") on node \"crc\" DevicePath \"\"" Sep 30 09:38:05 crc kubenswrapper[4810]: I0930 09:38:05.392907 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/crc-debug-92q5f" event={"ID":"2e93a8f1-6809-4d54-8231-cc351fe8d592","Type":"ContainerDied","Data":"65f6c9b960ec6ffb97573d1d7bdcc5acf3759a2e35f3d6e2eb9cd625de9f24af"} Sep 30 09:38:05 crc kubenswrapper[4810]: I0930 09:38:05.392967 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65f6c9b960ec6ffb97573d1d7bdcc5acf3759a2e35f3d6e2eb9cd625de9f24af" Sep 30 09:38:05 crc kubenswrapper[4810]: I0930 09:38:05.393039 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-92q5f" Sep 30 09:38:11 crc kubenswrapper[4810]: I0930 09:38:11.652941 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-48zlz/crc-debug-92q5f"] Sep 30 09:38:11 crc kubenswrapper[4810]: I0930 09:38:11.661302 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-48zlz/crc-debug-92q5f"] Sep 30 09:38:12 crc kubenswrapper[4810]: I0930 09:38:12.886285 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-48zlz/crc-debug-5cxx7"] Sep 30 09:38:12 crc kubenswrapper[4810]: E0930 09:38:12.887018 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e93a8f1-6809-4d54-8231-cc351fe8d592" containerName="container-00" Sep 30 09:38:12 crc kubenswrapper[4810]: I0930 09:38:12.887033 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e93a8f1-6809-4d54-8231-cc351fe8d592" containerName="container-00" Sep 30 09:38:12 crc kubenswrapper[4810]: I0930 09:38:12.887318 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e93a8f1-6809-4d54-8231-cc351fe8d592" containerName="container-00" Sep 30 09:38:12 crc kubenswrapper[4810]: I0930 09:38:12.888181 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-5cxx7" Sep 30 09:38:13 crc kubenswrapper[4810]: I0930 09:38:13.049323 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkjql\" (UniqueName: \"kubernetes.io/projected/f8252c4b-1d38-4368-905b-68cf60675703-kube-api-access-pkjql\") pod \"crc-debug-5cxx7\" (UID: \"f8252c4b-1d38-4368-905b-68cf60675703\") " pod="openshift-must-gather-48zlz/crc-debug-5cxx7" Sep 30 09:38:13 crc kubenswrapper[4810]: I0930 09:38:13.049521 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f8252c4b-1d38-4368-905b-68cf60675703-host\") pod \"crc-debug-5cxx7\" (UID: \"f8252c4b-1d38-4368-905b-68cf60675703\") " pod="openshift-must-gather-48zlz/crc-debug-5cxx7" Sep 30 09:38:13 crc kubenswrapper[4810]: I0930 09:38:13.150905 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f8252c4b-1d38-4368-905b-68cf60675703-host\") pod \"crc-debug-5cxx7\" (UID: \"f8252c4b-1d38-4368-905b-68cf60675703\") " pod="openshift-must-gather-48zlz/crc-debug-5cxx7" Sep 30 09:38:13 crc kubenswrapper[4810]: I0930 09:38:13.151036 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkjql\" (UniqueName: \"kubernetes.io/projected/f8252c4b-1d38-4368-905b-68cf60675703-kube-api-access-pkjql\") pod \"crc-debug-5cxx7\" (UID: \"f8252c4b-1d38-4368-905b-68cf60675703\") " pod="openshift-must-gather-48zlz/crc-debug-5cxx7" Sep 30 09:38:13 crc kubenswrapper[4810]: I0930 09:38:13.151055 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f8252c4b-1d38-4368-905b-68cf60675703-host\") pod \"crc-debug-5cxx7\" (UID: \"f8252c4b-1d38-4368-905b-68cf60675703\") " pod="openshift-must-gather-48zlz/crc-debug-5cxx7" Sep 30 09:38:13 crc kubenswrapper[4810]: I0930 09:38:13.174361 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkjql\" (UniqueName: \"kubernetes.io/projected/f8252c4b-1d38-4368-905b-68cf60675703-kube-api-access-pkjql\") pod \"crc-debug-5cxx7\" (UID: \"f8252c4b-1d38-4368-905b-68cf60675703\") " pod="openshift-must-gather-48zlz/crc-debug-5cxx7" Sep 30 09:38:13 crc kubenswrapper[4810]: I0930 09:38:13.203713 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-5cxx7" Sep 30 09:38:13 crc kubenswrapper[4810]: I0930 09:38:13.320462 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e93a8f1-6809-4d54-8231-cc351fe8d592" path="/var/lib/kubelet/pods/2e93a8f1-6809-4d54-8231-cc351fe8d592/volumes" Sep 30 09:38:13 crc kubenswrapper[4810]: I0930 09:38:13.469607 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/crc-debug-5cxx7" event={"ID":"f8252c4b-1d38-4368-905b-68cf60675703","Type":"ContainerStarted","Data":"394d8d4534ca3073072f7faadec74a63650e6acd9b07eabb4c8ebff41ac43fd0"} Sep 30 09:38:14 crc kubenswrapper[4810]: I0930 09:38:14.482653 4810 generic.go:334] "Generic (PLEG): container finished" podID="f8252c4b-1d38-4368-905b-68cf60675703" containerID="9e5dfbe73f6dd85b258d1e3aaebebdad03f92f184e7b315a52fe812fbaa9aafc" exitCode=0 Sep 30 09:38:14 crc kubenswrapper[4810]: I0930 09:38:14.482750 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/crc-debug-5cxx7" event={"ID":"f8252c4b-1d38-4368-905b-68cf60675703","Type":"ContainerDied","Data":"9e5dfbe73f6dd85b258d1e3aaebebdad03f92f184e7b315a52fe812fbaa9aafc"} Sep 30 09:38:14 crc kubenswrapper[4810]: I0930 09:38:14.521146 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-48zlz/crc-debug-5cxx7"] Sep 30 09:38:14 crc kubenswrapper[4810]: I0930 09:38:14.528752 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-48zlz/crc-debug-5cxx7"] Sep 30 09:38:15 crc kubenswrapper[4810]: I0930 09:38:15.627663 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-5cxx7" Sep 30 09:38:15 crc kubenswrapper[4810]: I0930 09:38:15.821846 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f8252c4b-1d38-4368-905b-68cf60675703-host\") pod \"f8252c4b-1d38-4368-905b-68cf60675703\" (UID: \"f8252c4b-1d38-4368-905b-68cf60675703\") " Sep 30 09:38:15 crc kubenswrapper[4810]: I0930 09:38:15.821918 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkjql\" (UniqueName: \"kubernetes.io/projected/f8252c4b-1d38-4368-905b-68cf60675703-kube-api-access-pkjql\") pod \"f8252c4b-1d38-4368-905b-68cf60675703\" (UID: \"f8252c4b-1d38-4368-905b-68cf60675703\") " Sep 30 09:38:15 crc kubenswrapper[4810]: I0930 09:38:15.821995 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8252c4b-1d38-4368-905b-68cf60675703-host" (OuterVolumeSpecName: "host") pod "f8252c4b-1d38-4368-905b-68cf60675703" (UID: "f8252c4b-1d38-4368-905b-68cf60675703"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 09:38:15 crc kubenswrapper[4810]: I0930 09:38:15.822399 4810 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f8252c4b-1d38-4368-905b-68cf60675703-host\") on node \"crc\" DevicePath \"\"" Sep 30 09:38:15 crc kubenswrapper[4810]: I0930 09:38:15.828529 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8252c4b-1d38-4368-905b-68cf60675703-kube-api-access-pkjql" (OuterVolumeSpecName: "kube-api-access-pkjql") pod "f8252c4b-1d38-4368-905b-68cf60675703" (UID: "f8252c4b-1d38-4368-905b-68cf60675703"). InnerVolumeSpecName "kube-api-access-pkjql". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:38:15 crc kubenswrapper[4810]: I0930 09:38:15.924367 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkjql\" (UniqueName: \"kubernetes.io/projected/f8252c4b-1d38-4368-905b-68cf60675703-kube-api-access-pkjql\") on node \"crc\" DevicePath \"\"" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.194413 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/util/0.log" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.349382 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/pull/0.log" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.360355 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/pull/0.log" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.378674 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/util/0.log" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.518431 4810 scope.go:117] "RemoveContainer" containerID="9e5dfbe73f6dd85b258d1e3aaebebdad03f92f184e7b315a52fe812fbaa9aafc" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.518583 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/crc-debug-5cxx7" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.606167 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/util/0.log" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.627732 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/extract/0.log" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.627863 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/pull/0.log" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.815846 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-zzj8m_6ad79408-8dc2-4e7e-aee3-54e3778a344e/kube-rbac-proxy/0.log" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.865203 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-zzj8m_6ad79408-8dc2-4e7e-aee3-54e3778a344e/manager/0.log" Sep 30 09:38:16 crc kubenswrapper[4810]: I0930 09:38:16.882638 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-6dhj8_982a5d6f-5897-40d0-88a4-40a9f1ae9a5f/kube-rbac-proxy/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.036747 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-6dhj8_982a5d6f-5897-40d0-88a4-40a9f1ae9a5f/manager/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.054971 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-zfcrn_02d77694-4bb0-4735-94c1-a3314ee634b9/kube-rbac-proxy/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.098961 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-zfcrn_02d77694-4bb0-4735-94c1-a3314ee634b9/manager/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.209687 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-9wfgs_da3d67ab-80be-486a-bc4d-d36ae6eb4756/kube-rbac-proxy/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.318282 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8252c4b-1d38-4368-905b-68cf60675703" path="/var/lib/kubelet/pods/f8252c4b-1d38-4368-905b-68cf60675703/volumes" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.321917 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-9wfgs_da3d67ab-80be-486a-bc4d-d36ae6eb4756/manager/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.410996 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-npvdf_e801f35d-5d9f-48df-9a66-dd11ca14d26c/kube-rbac-proxy/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.484013 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-npvdf_e801f35d-5d9f-48df-9a66-dd11ca14d26c/manager/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.542699 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-g76mw_09ab9971-634a-40dd-88df-c1093fb66766/kube-rbac-proxy/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.634714 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-g76mw_09ab9971-634a-40dd-88df-c1093fb66766/manager/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.726166 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-v9pjf_8b3a8007-4969-4c86-a03a-ce633b3c31aa/kube-rbac-proxy/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.868870 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-hd4x4_8c5f7678-0f63-4323-910c-227040fdac66/kube-rbac-proxy/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.914727 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-v9pjf_8b3a8007-4969-4c86-a03a-ce633b3c31aa/manager/0.log" Sep 30 09:38:17 crc kubenswrapper[4810]: I0930 09:38:17.972575 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-hd4x4_8c5f7678-0f63-4323-910c-227040fdac66/manager/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.040513 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-vz9pd_52a2d49f-0736-4cac-8363-1a71fa6c1dd4/kube-rbac-proxy/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.170694 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-vz9pd_52a2d49f-0736-4cac-8363-1a71fa6c1dd4/manager/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.286993 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-8bwwd_df615184-edc5-4d93-b0ae-b414d9eb0ce1/kube-rbac-proxy/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.287184 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-8bwwd_df615184-edc5-4d93-b0ae-b414d9eb0ce1/manager/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.383639 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-wnngc_8a8a969d-2c5d-4229-b01d-cabcf3aec35c/kube-rbac-proxy/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.497203 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-wnngc_8a8a969d-2c5d-4229-b01d-cabcf3aec35c/manager/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.567887 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-t6wwm_ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e/kube-rbac-proxy/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.629718 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-t6wwm_ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e/manager/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.706721 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-v58wt_b32751fb-95c0-4295-937a-ce7f82562863/kube-rbac-proxy/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.854733 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-v58wt_b32751fb-95c0-4295-937a-ce7f82562863/manager/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.912084 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-sndbm_0f34a7d1-aa5e-4b55-a75d-5366e196e742/manager/0.log" Sep 30 09:38:18 crc kubenswrapper[4810]: I0930 09:38:18.926121 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-sndbm_0f34a7d1-aa5e-4b55-a75d-5366e196e742/kube-rbac-proxy/0.log" Sep 30 09:38:19 crc kubenswrapper[4810]: I0930 09:38:19.103527 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-9gsjq_4e9c80fa-52fc-41db-a0be-babb0bd14b56/manager/0.log" Sep 30 09:38:19 crc kubenswrapper[4810]: I0930 09:38:19.108566 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-9gsjq_4e9c80fa-52fc-41db-a0be-babb0bd14b56/kube-rbac-proxy/0.log" Sep 30 09:38:19 crc kubenswrapper[4810]: I0930 09:38:19.223629 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-64647d55fc-6z69x_3df3a4ac-fc49-404f-aaba-cc8c053365c4/kube-rbac-proxy/0.log" Sep 30 09:38:19 crc kubenswrapper[4810]: I0930 09:38:19.397303 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6d66d9d974-d8fdr_154d7384-4a63-4a16-8400-903c4fe9eb8b/kube-rbac-proxy/0.log" Sep 30 09:38:19 crc kubenswrapper[4810]: I0930 09:38:19.600575 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-fqg48_75091a5f-7967-4317-90b6-e195f197ecdb/registry-server/0.log" Sep 30 09:38:19 crc kubenswrapper[4810]: I0930 09:38:19.648207 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6d66d9d974-d8fdr_154d7384-4a63-4a16-8400-903c4fe9eb8b/operator/0.log" Sep 30 09:38:19 crc kubenswrapper[4810]: I0930 09:38:19.943721 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-s785j_24044399-d033-436a-ad32-1e1aefe76626/kube-rbac-proxy/0.log" Sep 30 09:38:19 crc kubenswrapper[4810]: I0930 09:38:19.995434 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-s785j_24044399-d033-436a-ad32-1e1aefe76626/manager/0.log" Sep 30 09:38:20 crc kubenswrapper[4810]: I0930 09:38:20.143873 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-t24hl_11f06c30-1cef-40f3-a493-8be235930b71/kube-rbac-proxy/0.log" Sep 30 09:38:20 crc kubenswrapper[4810]: I0930 09:38:20.244106 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-t24hl_11f06c30-1cef-40f3-a493-8be235930b71/manager/0.log" Sep 30 09:38:20 crc kubenswrapper[4810]: I0930 09:38:20.329837 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-79d8469568-js8pw_bf4551e1-3981-407f-ba07-bff4db2b6d6d/operator/0.log" Sep 30 09:38:20 crc kubenswrapper[4810]: I0930 09:38:20.539893 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-64647d55fc-6z69x_3df3a4ac-fc49-404f-aaba-cc8c053365c4/manager/0.log" Sep 30 09:38:20 crc kubenswrapper[4810]: I0930 09:38:20.631879 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-tc2wc_3bbbc1cd-370d-468f-8985-cdf9219ca326/kube-rbac-proxy/0.log" Sep 30 09:38:20 crc kubenswrapper[4810]: I0930 09:38:20.721919 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-tc2wc_3bbbc1cd-370d-468f-8985-cdf9219ca326/manager/0.log" Sep 30 09:38:20 crc kubenswrapper[4810]: I0930 09:38:20.833992 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-g8ls7_0cc20b2d-4504-4f46-b0a0-5aee4b2201a3/kube-rbac-proxy/0.log" Sep 30 09:38:20 crc kubenswrapper[4810]: I0930 09:38:20.979916 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-g8s55_7fe8706c-6dfa-4a2f-876b-4cded24209bf/kube-rbac-proxy/0.log" Sep 30 09:38:21 crc kubenswrapper[4810]: I0930 09:38:21.054794 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-g8s55_7fe8706c-6dfa-4a2f-876b-4cded24209bf/manager/0.log" Sep 30 09:38:21 crc kubenswrapper[4810]: I0930 09:38:21.104109 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-59446d9d9-kb958_b63222b9-1cc3-4a55-a498-ce538cea80b9/kube-rbac-proxy/0.log" Sep 30 09:38:21 crc kubenswrapper[4810]: I0930 09:38:21.116089 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-g8ls7_0cc20b2d-4504-4f46-b0a0-5aee4b2201a3/manager/0.log" Sep 30 09:38:21 crc kubenswrapper[4810]: I0930 09:38:21.290714 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-59446d9d9-kb958_b63222b9-1cc3-4a55-a498-ce538cea80b9/manager/0.log" Sep 30 09:38:37 crc kubenswrapper[4810]: I0930 09:38:37.327129 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-p7cc4_00e1ebdd-3664-401d-b7bf-a9f786d9b6d2/control-plane-machine-set-operator/0.log" Sep 30 09:38:37 crc kubenswrapper[4810]: I0930 09:38:37.469171 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nj7lz_f7201c6b-de6b-4c4e-b20b-90c59d0db7ac/kube-rbac-proxy/0.log" Sep 30 09:38:37 crc kubenswrapper[4810]: I0930 09:38:37.514754 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nj7lz_f7201c6b-de6b-4c4e-b20b-90c59d0db7ac/machine-api-operator/0.log" Sep 30 09:38:45 crc kubenswrapper[4810]: I0930 09:38:45.911498 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:38:45 crc kubenswrapper[4810]: I0930 09:38:45.912074 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:38:48 crc kubenswrapper[4810]: I0930 09:38:48.765296 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-bdbrw_14ff18f7-0dac-4905-a94e-5a687ad05c8a/cert-manager-controller/0.log" Sep 30 09:38:48 crc kubenswrapper[4810]: I0930 09:38:48.978344 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-wdstp_8090d644-f567-428e-a3b4-0d529d9bccb9/cert-manager-webhook/0.log" Sep 30 09:38:48 crc kubenswrapper[4810]: I0930 09:38:48.983765 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-vbt4m_28b9d5cb-a422-4629-ba55-8bcd92f69971/cert-manager-cainjector/0.log" Sep 30 09:39:00 crc kubenswrapper[4810]: I0930 09:39:00.162600 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-pdhp7_b672b0bf-c5cc-467a-9014-99a3269488e8/nmstate-console-plugin/0.log" Sep 30 09:39:00 crc kubenswrapper[4810]: I0930 09:39:00.321998 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-w4dnr_b8df67fa-d671-4174-8306-61127ae22268/nmstate-handler/0.log" Sep 30 09:39:00 crc kubenswrapper[4810]: I0930 09:39:00.392456 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-qzvw4_303e8028-1bcf-4fa7-8314-2ed52bec230a/kube-rbac-proxy/0.log" Sep 30 09:39:00 crc kubenswrapper[4810]: I0930 09:39:00.400042 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-qzvw4_303e8028-1bcf-4fa7-8314-2ed52bec230a/nmstate-metrics/0.log" Sep 30 09:39:00 crc kubenswrapper[4810]: I0930 09:39:00.534493 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-bk866_5a1ec2d4-00e6-4db8-95c0-3079d24f18a8/nmstate-operator/0.log" Sep 30 09:39:00 crc kubenswrapper[4810]: I0930 09:39:00.570848 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-h2zdm_8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb/nmstate-webhook/0.log" Sep 30 09:39:09 crc kubenswrapper[4810]: I0930 09:39:09.934822 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w9rh8"] Sep 30 09:39:09 crc kubenswrapper[4810]: E0930 09:39:09.935906 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8252c4b-1d38-4368-905b-68cf60675703" containerName="container-00" Sep 30 09:39:09 crc kubenswrapper[4810]: I0930 09:39:09.935922 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8252c4b-1d38-4368-905b-68cf60675703" containerName="container-00" Sep 30 09:39:09 crc kubenswrapper[4810]: I0930 09:39:09.936156 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8252c4b-1d38-4368-905b-68cf60675703" containerName="container-00" Sep 30 09:39:09 crc kubenswrapper[4810]: I0930 09:39:09.937773 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:09 crc kubenswrapper[4810]: I0930 09:39:09.948896 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9rh8"] Sep 30 09:39:09 crc kubenswrapper[4810]: I0930 09:39:09.984510 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-catalog-content\") pod \"redhat-marketplace-w9rh8\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:09 crc kubenswrapper[4810]: I0930 09:39:09.984854 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwnx6\" (UniqueName: \"kubernetes.io/projected/3adf5e65-a297-40af-b1e6-b7fa0af1d779-kube-api-access-rwnx6\") pod \"redhat-marketplace-w9rh8\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:09 crc kubenswrapper[4810]: I0930 09:39:09.985032 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-utilities\") pod \"redhat-marketplace-w9rh8\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:10 crc kubenswrapper[4810]: I0930 09:39:10.086578 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-utilities\") pod \"redhat-marketplace-w9rh8\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:10 crc kubenswrapper[4810]: I0930 09:39:10.086670 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-catalog-content\") pod \"redhat-marketplace-w9rh8\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:10 crc kubenswrapper[4810]: I0930 09:39:10.087364 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-catalog-content\") pod \"redhat-marketplace-w9rh8\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:10 crc kubenswrapper[4810]: I0930 09:39:10.087368 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-utilities\") pod \"redhat-marketplace-w9rh8\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:10 crc kubenswrapper[4810]: I0930 09:39:10.087520 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwnx6\" (UniqueName: \"kubernetes.io/projected/3adf5e65-a297-40af-b1e6-b7fa0af1d779-kube-api-access-rwnx6\") pod \"redhat-marketplace-w9rh8\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:10 crc kubenswrapper[4810]: I0930 09:39:10.115198 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwnx6\" (UniqueName: \"kubernetes.io/projected/3adf5e65-a297-40af-b1e6-b7fa0af1d779-kube-api-access-rwnx6\") pod \"redhat-marketplace-w9rh8\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:10 crc kubenswrapper[4810]: I0930 09:39:10.272700 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:10 crc kubenswrapper[4810]: I0930 09:39:10.783069 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9rh8"] Sep 30 09:39:11 crc kubenswrapper[4810]: I0930 09:39:11.040115 4810 generic.go:334] "Generic (PLEG): container finished" podID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" containerID="3c6bf6369a541da2029802c1eca3368c7bd405d5b04b609509f07b3003ae6897" exitCode=0 Sep 30 09:39:11 crc kubenswrapper[4810]: I0930 09:39:11.040171 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9rh8" event={"ID":"3adf5e65-a297-40af-b1e6-b7fa0af1d779","Type":"ContainerDied","Data":"3c6bf6369a541da2029802c1eca3368c7bd405d5b04b609509f07b3003ae6897"} Sep 30 09:39:11 crc kubenswrapper[4810]: I0930 09:39:11.040203 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9rh8" event={"ID":"3adf5e65-a297-40af-b1e6-b7fa0af1d779","Type":"ContainerStarted","Data":"01d11cdca9439e736a688a64af4ae18b98ccaff3d2315b83a3e8d61eabefb795"} Sep 30 09:39:12 crc kubenswrapper[4810]: I0930 09:39:12.051423 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9rh8" event={"ID":"3adf5e65-a297-40af-b1e6-b7fa0af1d779","Type":"ContainerStarted","Data":"1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393"} Sep 30 09:39:13 crc kubenswrapper[4810]: I0930 09:39:13.063826 4810 generic.go:334] "Generic (PLEG): container finished" podID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" containerID="1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393" exitCode=0 Sep 30 09:39:13 crc kubenswrapper[4810]: I0930 09:39:13.063909 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9rh8" event={"ID":"3adf5e65-a297-40af-b1e6-b7fa0af1d779","Type":"ContainerDied","Data":"1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393"} Sep 30 09:39:13 crc kubenswrapper[4810]: I0930 09:39:13.657931 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-qr7xw_4f9b2513-5122-4757-b48c-4f0e8c9aaf05/kube-rbac-proxy/0.log" Sep 30 09:39:13 crc kubenswrapper[4810]: I0930 09:39:13.812587 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-qr7xw_4f9b2513-5122-4757-b48c-4f0e8c9aaf05/controller/0.log" Sep 30 09:39:13 crc kubenswrapper[4810]: I0930 09:39:13.877860 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-2cfm4_ad0bc482-465c-433b-bd10-82ed2f79b6ad/frr-k8s-webhook-server/0.log" Sep 30 09:39:13 crc kubenswrapper[4810]: I0930 09:39:13.982495 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-frr-files/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.076539 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9rh8" event={"ID":"3adf5e65-a297-40af-b1e6-b7fa0af1d779","Type":"ContainerStarted","Data":"7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0"} Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.096353 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w9rh8" podStartSLOduration=2.610901238 podStartE2EDuration="5.09633211s" podCreationTimestamp="2025-09-30 09:39:09 +0000 UTC" firstStartedPulling="2025-09-30 09:39:11.042728636 +0000 UTC m=+5774.494927903" lastFinishedPulling="2025-09-30 09:39:13.528159498 +0000 UTC m=+5776.980358775" observedRunningTime="2025-09-30 09:39:14.095381253 +0000 UTC m=+5777.547580520" watchObservedRunningTime="2025-09-30 09:39:14.09633211 +0000 UTC m=+5777.548531377" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.191433 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-reloader/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.230767 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-reloader/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.247190 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-metrics/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.254283 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-frr-files/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.432170 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-frr-files/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.498286 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-metrics/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.513402 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-reloader/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.520581 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-metrics/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.654381 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-frr-files/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.660766 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-metrics/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.699378 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/controller/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.703623 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-reloader/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.813515 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/frr-metrics/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.863679 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/kube-rbac-proxy/0.log" Sep 30 09:39:14 crc kubenswrapper[4810]: I0930 09:39:14.899004 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/kube-rbac-proxy-frr/0.log" Sep 30 09:39:15 crc kubenswrapper[4810]: I0930 09:39:15.051240 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/reloader/0.log" Sep 30 09:39:15 crc kubenswrapper[4810]: I0930 09:39:15.135018 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-74b66975c5-rtrd5_9710139d-e7f3-4921-b84a-4b6525891f99/manager/0.log" Sep 30 09:39:15 crc kubenswrapper[4810]: I0930 09:39:15.345966 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-768c8b4c95-d2v27_9d49935b-5628-43f0-b527-18d3b55aaa3a/webhook-server/0.log" Sep 30 09:39:15 crc kubenswrapper[4810]: I0930 09:39:15.473195 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-gjfhs_2f29bb1f-4702-4fa8-8a9e-692047b356f9/kube-rbac-proxy/0.log" Sep 30 09:39:15 crc kubenswrapper[4810]: I0930 09:39:15.910966 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:39:15 crc kubenswrapper[4810]: I0930 09:39:15.911014 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:39:16 crc kubenswrapper[4810]: I0930 09:39:16.085151 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-gjfhs_2f29bb1f-4702-4fa8-8a9e-692047b356f9/speaker/0.log" Sep 30 09:39:16 crc kubenswrapper[4810]: I0930 09:39:16.525438 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/frr/0.log" Sep 30 09:39:20 crc kubenswrapper[4810]: I0930 09:39:20.273314 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:20 crc kubenswrapper[4810]: I0930 09:39:20.273653 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:20 crc kubenswrapper[4810]: I0930 09:39:20.332757 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:21 crc kubenswrapper[4810]: I0930 09:39:21.205591 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:21 crc kubenswrapper[4810]: I0930 09:39:21.254772 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9rh8"] Sep 30 09:39:23 crc kubenswrapper[4810]: I0930 09:39:23.168081 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w9rh8" podUID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" containerName="registry-server" containerID="cri-o://7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0" gracePeriod=2 Sep 30 09:39:23 crc kubenswrapper[4810]: I0930 09:39:23.681985 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:23 crc kubenswrapper[4810]: I0930 09:39:23.750104 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwnx6\" (UniqueName: \"kubernetes.io/projected/3adf5e65-a297-40af-b1e6-b7fa0af1d779-kube-api-access-rwnx6\") pod \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " Sep 30 09:39:23 crc kubenswrapper[4810]: I0930 09:39:23.750308 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-utilities\") pod \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " Sep 30 09:39:23 crc kubenswrapper[4810]: I0930 09:39:23.750608 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-catalog-content\") pod \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\" (UID: \"3adf5e65-a297-40af-b1e6-b7fa0af1d779\") " Sep 30 09:39:23 crc kubenswrapper[4810]: I0930 09:39:23.754239 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-utilities" (OuterVolumeSpecName: "utilities") pod "3adf5e65-a297-40af-b1e6-b7fa0af1d779" (UID: "3adf5e65-a297-40af-b1e6-b7fa0af1d779"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:39:23 crc kubenswrapper[4810]: I0930 09:39:23.759491 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3adf5e65-a297-40af-b1e6-b7fa0af1d779-kube-api-access-rwnx6" (OuterVolumeSpecName: "kube-api-access-rwnx6") pod "3adf5e65-a297-40af-b1e6-b7fa0af1d779" (UID: "3adf5e65-a297-40af-b1e6-b7fa0af1d779"). InnerVolumeSpecName "kube-api-access-rwnx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:39:23 crc kubenswrapper[4810]: I0930 09:39:23.779876 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3adf5e65-a297-40af-b1e6-b7fa0af1d779" (UID: "3adf5e65-a297-40af-b1e6-b7fa0af1d779"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:39:23 crc kubenswrapper[4810]: I0930 09:39:23.853340 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:39:23 crc kubenswrapper[4810]: I0930 09:39:23.853415 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwnx6\" (UniqueName: \"kubernetes.io/projected/3adf5e65-a297-40af-b1e6-b7fa0af1d779-kube-api-access-rwnx6\") on node \"crc\" DevicePath \"\"" Sep 30 09:39:23 crc kubenswrapper[4810]: I0930 09:39:23.853445 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adf5e65-a297-40af-b1e6-b7fa0af1d779-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.178043 4810 generic.go:334] "Generic (PLEG): container finished" podID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" containerID="7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0" exitCode=0 Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.178133 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w9rh8" Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.178141 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9rh8" event={"ID":"3adf5e65-a297-40af-b1e6-b7fa0af1d779","Type":"ContainerDied","Data":"7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0"} Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.179138 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9rh8" event={"ID":"3adf5e65-a297-40af-b1e6-b7fa0af1d779","Type":"ContainerDied","Data":"01d11cdca9439e736a688a64af4ae18b98ccaff3d2315b83a3e8d61eabefb795"} Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.179157 4810 scope.go:117] "RemoveContainer" containerID="7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0" Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.197475 4810 scope.go:117] "RemoveContainer" containerID="1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393" Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.216819 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9rh8"] Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.225067 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9rh8"] Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.233618 4810 scope.go:117] "RemoveContainer" containerID="3c6bf6369a541da2029802c1eca3368c7bd405d5b04b609509f07b3003ae6897" Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.272951 4810 scope.go:117] "RemoveContainer" containerID="7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0" Sep 30 09:39:24 crc kubenswrapper[4810]: E0930 09:39:24.273653 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0\": container with ID starting with 7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0 not found: ID does not exist" containerID="7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0" Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.273719 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0"} err="failed to get container status \"7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0\": rpc error: code = NotFound desc = could not find container \"7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0\": container with ID starting with 7ee5973c05a314f33791085d0283ad4b947dba98593f5f9f4b570764ef1e29f0 not found: ID does not exist" Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.273761 4810 scope.go:117] "RemoveContainer" containerID="1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393" Sep 30 09:39:24 crc kubenswrapper[4810]: E0930 09:39:24.274303 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393\": container with ID starting with 1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393 not found: ID does not exist" containerID="1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393" Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.274346 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393"} err="failed to get container status \"1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393\": rpc error: code = NotFound desc = could not find container \"1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393\": container with ID starting with 1251026467f01a3aec5d95c122f9f8af062147ba90a2a4116c1890dba2555393 not found: ID does not exist" Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.274375 4810 scope.go:117] "RemoveContainer" containerID="3c6bf6369a541da2029802c1eca3368c7bd405d5b04b609509f07b3003ae6897" Sep 30 09:39:24 crc kubenswrapper[4810]: E0930 09:39:24.274699 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c6bf6369a541da2029802c1eca3368c7bd405d5b04b609509f07b3003ae6897\": container with ID starting with 3c6bf6369a541da2029802c1eca3368c7bd405d5b04b609509f07b3003ae6897 not found: ID does not exist" containerID="3c6bf6369a541da2029802c1eca3368c7bd405d5b04b609509f07b3003ae6897" Sep 30 09:39:24 crc kubenswrapper[4810]: I0930 09:39:24.274721 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c6bf6369a541da2029802c1eca3368c7bd405d5b04b609509f07b3003ae6897"} err="failed to get container status \"3c6bf6369a541da2029802c1eca3368c7bd405d5b04b609509f07b3003ae6897\": rpc error: code = NotFound desc = could not find container \"3c6bf6369a541da2029802c1eca3368c7bd405d5b04b609509f07b3003ae6897\": container with ID starting with 3c6bf6369a541da2029802c1eca3368c7bd405d5b04b609509f07b3003ae6897 not found: ID does not exist" Sep 30 09:39:25 crc kubenswrapper[4810]: I0930 09:39:25.317543 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" path="/var/lib/kubelet/pods/3adf5e65-a297-40af-b1e6-b7fa0af1d779/volumes" Sep 30 09:39:27 crc kubenswrapper[4810]: I0930 09:39:27.706516 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/util/0.log" Sep 30 09:39:27 crc kubenswrapper[4810]: I0930 09:39:27.857221 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/util/0.log" Sep 30 09:39:27 crc kubenswrapper[4810]: I0930 09:39:27.882854 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/pull/0.log" Sep 30 09:39:27 crc kubenswrapper[4810]: I0930 09:39:27.886738 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/pull/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.011649 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/util/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.040388 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/pull/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.080559 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/extract/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.191916 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/util/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.381151 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/pull/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.389561 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/util/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.410667 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/pull/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.565662 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/util/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.570814 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/pull/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.581485 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/extract/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.756167 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-utilities/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.913972 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-content/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.961967 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-content/0.log" Sep 30 09:39:28 crc kubenswrapper[4810]: I0930 09:39:28.967350 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-utilities/0.log" Sep 30 09:39:29 crc kubenswrapper[4810]: I0930 09:39:29.137233 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-content/0.log" Sep 30 09:39:29 crc kubenswrapper[4810]: I0930 09:39:29.151594 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-utilities/0.log" Sep 30 09:39:29 crc kubenswrapper[4810]: I0930 09:39:29.327599 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-utilities/0.log" Sep 30 09:39:29 crc kubenswrapper[4810]: I0930 09:39:29.490346 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-utilities/0.log" Sep 30 09:39:29 crc kubenswrapper[4810]: I0930 09:39:29.558418 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-content/0.log" Sep 30 09:39:29 crc kubenswrapper[4810]: I0930 09:39:29.574474 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-content/0.log" Sep 30 09:39:29 crc kubenswrapper[4810]: I0930 09:39:29.767572 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-utilities/0.log" Sep 30 09:39:29 crc kubenswrapper[4810]: I0930 09:39:29.861379 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-content/0.log" Sep 30 09:39:30 crc kubenswrapper[4810]: I0930 09:39:30.121039 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/util/0.log" Sep 30 09:39:30 crc kubenswrapper[4810]: I0930 09:39:30.386739 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/pull/0.log" Sep 30 09:39:30 crc kubenswrapper[4810]: I0930 09:39:30.413155 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/util/0.log" Sep 30 09:39:30 crc kubenswrapper[4810]: I0930 09:39:30.444663 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/registry-server/0.log" Sep 30 09:39:30 crc kubenswrapper[4810]: I0930 09:39:30.572128 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/pull/0.log" Sep 30 09:39:30 crc kubenswrapper[4810]: I0930 09:39:30.631580 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/registry-server/0.log" Sep 30 09:39:30 crc kubenswrapper[4810]: I0930 09:39:30.821626 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/pull/0.log" Sep 30 09:39:30 crc kubenswrapper[4810]: I0930 09:39:30.846036 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/util/0.log" Sep 30 09:39:30 crc kubenswrapper[4810]: I0930 09:39:30.850245 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/extract/0.log" Sep 30 09:39:30 crc kubenswrapper[4810]: I0930 09:39:30.985701 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-txqd2_7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba/marketplace-operator/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.018727 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-utilities/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.203923 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-content/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.230251 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-utilities/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.231994 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-content/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.403785 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-content/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.416650 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-utilities/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.451708 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-utilities/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.593876 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/registry-server/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.641796 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-content/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.672082 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-utilities/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.674137 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-content/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.830170 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-content/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.863518 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-utilities/0.log" Sep 30 09:39:31 crc kubenswrapper[4810]: I0930 09:39:31.986188 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/registry-server/0.log" Sep 30 09:39:43 crc kubenswrapper[4810]: I0930 09:39:43.587377 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-9cddq_360a4c93-d93b-4caf-b11d-a2185a5179d3/prometheus-operator/0.log" Sep 30 09:39:43 crc kubenswrapper[4810]: I0930 09:39:43.721348 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p_e41015ca-c417-4076-9895-abff0488bfd7/prometheus-operator-admission-webhook/0.log" Sep 30 09:39:43 crc kubenswrapper[4810]: I0930 09:39:43.782700 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8_2ea01519-d717-45ef-be26-b429eb2c6cdb/prometheus-operator-admission-webhook/0.log" Sep 30 09:39:43 crc kubenswrapper[4810]: I0930 09:39:43.929565 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-nq2k6_0da3209d-75c8-4eda-97f4-b6cfeac48f62/operator/0.log" Sep 30 09:39:43 crc kubenswrapper[4810]: I0930 09:39:43.962402 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-kj7xb_5f51c1dc-ac89-4a84-8897-d2e9770baec9/perses-operator/0.log" Sep 30 09:39:45 crc kubenswrapper[4810]: I0930 09:39:45.925141 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:39:45 crc kubenswrapper[4810]: I0930 09:39:45.925487 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:39:45 crc kubenswrapper[4810]: I0930 09:39:45.927120 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 09:39:45 crc kubenswrapper[4810]: I0930 09:39:45.928427 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3cfe1e0b8905c291915720129de0e23bdd06ea6549e1c8ca139a56a2878a94dd"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 09:39:45 crc kubenswrapper[4810]: I0930 09:39:45.928487 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://3cfe1e0b8905c291915720129de0e23bdd06ea6549e1c8ca139a56a2878a94dd" gracePeriod=600 Sep 30 09:39:46 crc kubenswrapper[4810]: I0930 09:39:46.458961 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="3cfe1e0b8905c291915720129de0e23bdd06ea6549e1c8ca139a56a2878a94dd" exitCode=0 Sep 30 09:39:46 crc kubenswrapper[4810]: I0930 09:39:46.459327 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"3cfe1e0b8905c291915720129de0e23bdd06ea6549e1c8ca139a56a2878a94dd"} Sep 30 09:39:46 crc kubenswrapper[4810]: I0930 09:39:46.459782 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e"} Sep 30 09:39:46 crc kubenswrapper[4810]: I0930 09:39:46.459816 4810 scope.go:117] "RemoveContainer" containerID="d9064f16ca65f4854dc425145390bd3787462666a9ce0a878d0c797410044c68" Sep 30 09:40:04 crc kubenswrapper[4810]: E0930 09:40:04.041750 4810 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.203:34428->38.102.83.203:42869: write tcp 38.102.83.203:34428->38.102.83.203:42869: write: broken pipe Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.281996 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n5cvg"] Sep 30 09:41:16 crc kubenswrapper[4810]: E0930 09:41:16.282871 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" containerName="registry-server" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.282885 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" containerName="registry-server" Sep 30 09:41:16 crc kubenswrapper[4810]: E0930 09:41:16.282915 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" containerName="extract-utilities" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.282922 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" containerName="extract-utilities" Sep 30 09:41:16 crc kubenswrapper[4810]: E0930 09:41:16.282936 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" containerName="extract-content" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.282943 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" containerName="extract-content" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.283135 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3adf5e65-a297-40af-b1e6-b7fa0af1d779" containerName="registry-server" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.284514 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.324388 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-utilities\") pod \"certified-operators-n5cvg\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.324427 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wkvx\" (UniqueName: \"kubernetes.io/projected/1167049c-b656-4f5e-ba9f-ae8247a60752-kube-api-access-8wkvx\") pod \"certified-operators-n5cvg\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.324443 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-catalog-content\") pod \"certified-operators-n5cvg\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.335544 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n5cvg"] Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.426296 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-utilities\") pod \"certified-operators-n5cvg\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.426353 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wkvx\" (UniqueName: \"kubernetes.io/projected/1167049c-b656-4f5e-ba9f-ae8247a60752-kube-api-access-8wkvx\") pod \"certified-operators-n5cvg\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.426382 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-catalog-content\") pod \"certified-operators-n5cvg\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.426803 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-utilities\") pod \"certified-operators-n5cvg\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.429612 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-catalog-content\") pod \"certified-operators-n5cvg\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.454378 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wkvx\" (UniqueName: \"kubernetes.io/projected/1167049c-b656-4f5e-ba9f-ae8247a60752-kube-api-access-8wkvx\") pod \"certified-operators-n5cvg\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:16 crc kubenswrapper[4810]: I0930 09:41:16.617334 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:17 crc kubenswrapper[4810]: I0930 09:41:17.128800 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n5cvg"] Sep 30 09:41:17 crc kubenswrapper[4810]: I0930 09:41:17.560537 4810 generic.go:334] "Generic (PLEG): container finished" podID="1167049c-b656-4f5e-ba9f-ae8247a60752" containerID="cf2578633a5ef11aa2429fe4ead736eecb0661384b0272f1d7817d65d2c0728b" exitCode=0 Sep 30 09:41:17 crc kubenswrapper[4810]: I0930 09:41:17.560603 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n5cvg" event={"ID":"1167049c-b656-4f5e-ba9f-ae8247a60752","Type":"ContainerDied","Data":"cf2578633a5ef11aa2429fe4ead736eecb0661384b0272f1d7817d65d2c0728b"} Sep 30 09:41:17 crc kubenswrapper[4810]: I0930 09:41:17.560642 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n5cvg" event={"ID":"1167049c-b656-4f5e-ba9f-ae8247a60752","Type":"ContainerStarted","Data":"a1d8e8ede0f91ad942c8915a9793131f33c634219f480b2f88255fcfcbe0484b"} Sep 30 09:41:17 crc kubenswrapper[4810]: I0930 09:41:17.563074 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 09:41:19 crc kubenswrapper[4810]: I0930 09:41:19.584328 4810 generic.go:334] "Generic (PLEG): container finished" podID="1167049c-b656-4f5e-ba9f-ae8247a60752" containerID="55688107c76f56ee24576bdca18f1de648960cdeb6dbe1ca6720cb453d481246" exitCode=0 Sep 30 09:41:19 crc kubenswrapper[4810]: I0930 09:41:19.584685 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n5cvg" event={"ID":"1167049c-b656-4f5e-ba9f-ae8247a60752","Type":"ContainerDied","Data":"55688107c76f56ee24576bdca18f1de648960cdeb6dbe1ca6720cb453d481246"} Sep 30 09:41:20 crc kubenswrapper[4810]: I0930 09:41:20.597671 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n5cvg" event={"ID":"1167049c-b656-4f5e-ba9f-ae8247a60752","Type":"ContainerStarted","Data":"8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd"} Sep 30 09:41:20 crc kubenswrapper[4810]: I0930 09:41:20.624481 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n5cvg" podStartSLOduration=1.8251493060000001 podStartE2EDuration="4.624459886s" podCreationTimestamp="2025-09-30 09:41:16 +0000 UTC" firstStartedPulling="2025-09-30 09:41:17.562689704 +0000 UTC m=+5901.014889011" lastFinishedPulling="2025-09-30 09:41:20.362000324 +0000 UTC m=+5903.814199591" observedRunningTime="2025-09-30 09:41:20.619398939 +0000 UTC m=+5904.071598216" watchObservedRunningTime="2025-09-30 09:41:20.624459886 +0000 UTC m=+5904.076659153" Sep 30 09:41:26 crc kubenswrapper[4810]: I0930 09:41:26.618249 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:26 crc kubenswrapper[4810]: I0930 09:41:26.618981 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:26 crc kubenswrapper[4810]: I0930 09:41:26.672466 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:26 crc kubenswrapper[4810]: I0930 09:41:26.724115 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:26 crc kubenswrapper[4810]: I0930 09:41:26.911729 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n5cvg"] Sep 30 09:41:28 crc kubenswrapper[4810]: I0930 09:41:28.686226 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n5cvg" podUID="1167049c-b656-4f5e-ba9f-ae8247a60752" containerName="registry-server" containerID="cri-o://8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd" gracePeriod=2 Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.151013 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.292358 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wkvx\" (UniqueName: \"kubernetes.io/projected/1167049c-b656-4f5e-ba9f-ae8247a60752-kube-api-access-8wkvx\") pod \"1167049c-b656-4f5e-ba9f-ae8247a60752\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.292486 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-catalog-content\") pod \"1167049c-b656-4f5e-ba9f-ae8247a60752\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.300207 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1167049c-b656-4f5e-ba9f-ae8247a60752-kube-api-access-8wkvx" (OuterVolumeSpecName: "kube-api-access-8wkvx") pod "1167049c-b656-4f5e-ba9f-ae8247a60752" (UID: "1167049c-b656-4f5e-ba9f-ae8247a60752"). InnerVolumeSpecName "kube-api-access-8wkvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.302606 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-utilities\") pod \"1167049c-b656-4f5e-ba9f-ae8247a60752\" (UID: \"1167049c-b656-4f5e-ba9f-ae8247a60752\") " Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.303639 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-utilities" (OuterVolumeSpecName: "utilities") pod "1167049c-b656-4f5e-ba9f-ae8247a60752" (UID: "1167049c-b656-4f5e-ba9f-ae8247a60752"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.304143 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wkvx\" (UniqueName: \"kubernetes.io/projected/1167049c-b656-4f5e-ba9f-ae8247a60752-kube-api-access-8wkvx\") on node \"crc\" DevicePath \"\"" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.304173 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.727842 4810 generic.go:334] "Generic (PLEG): container finished" podID="1167049c-b656-4f5e-ba9f-ae8247a60752" containerID="8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd" exitCode=0 Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.728025 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n5cvg" event={"ID":"1167049c-b656-4f5e-ba9f-ae8247a60752","Type":"ContainerDied","Data":"8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd"} Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.728470 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n5cvg" event={"ID":"1167049c-b656-4f5e-ba9f-ae8247a60752","Type":"ContainerDied","Data":"a1d8e8ede0f91ad942c8915a9793131f33c634219f480b2f88255fcfcbe0484b"} Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.728523 4810 scope.go:117] "RemoveContainer" containerID="8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.728098 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n5cvg" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.755872 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1167049c-b656-4f5e-ba9f-ae8247a60752" (UID: "1167049c-b656-4f5e-ba9f-ae8247a60752"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.761262 4810 scope.go:117] "RemoveContainer" containerID="55688107c76f56ee24576bdca18f1de648960cdeb6dbe1ca6720cb453d481246" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.789595 4810 scope.go:117] "RemoveContainer" containerID="cf2578633a5ef11aa2429fe4ead736eecb0661384b0272f1d7817d65d2c0728b" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.821589 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1167049c-b656-4f5e-ba9f-ae8247a60752-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.839087 4810 scope.go:117] "RemoveContainer" containerID="8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd" Sep 30 09:41:29 crc kubenswrapper[4810]: E0930 09:41:29.839636 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd\": container with ID starting with 8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd not found: ID does not exist" containerID="8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.839687 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd"} err="failed to get container status \"8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd\": rpc error: code = NotFound desc = could not find container \"8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd\": container with ID starting with 8f92da69c9f2f2b71ab237a895965837360310bb105befc802ffbf3ae8b555bd not found: ID does not exist" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.839720 4810 scope.go:117] "RemoveContainer" containerID="55688107c76f56ee24576bdca18f1de648960cdeb6dbe1ca6720cb453d481246" Sep 30 09:41:29 crc kubenswrapper[4810]: E0930 09:41:29.840164 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55688107c76f56ee24576bdca18f1de648960cdeb6dbe1ca6720cb453d481246\": container with ID starting with 55688107c76f56ee24576bdca18f1de648960cdeb6dbe1ca6720cb453d481246 not found: ID does not exist" containerID="55688107c76f56ee24576bdca18f1de648960cdeb6dbe1ca6720cb453d481246" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.840247 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55688107c76f56ee24576bdca18f1de648960cdeb6dbe1ca6720cb453d481246"} err="failed to get container status \"55688107c76f56ee24576bdca18f1de648960cdeb6dbe1ca6720cb453d481246\": rpc error: code = NotFound desc = could not find container \"55688107c76f56ee24576bdca18f1de648960cdeb6dbe1ca6720cb453d481246\": container with ID starting with 55688107c76f56ee24576bdca18f1de648960cdeb6dbe1ca6720cb453d481246 not found: ID does not exist" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.840365 4810 scope.go:117] "RemoveContainer" containerID="cf2578633a5ef11aa2429fe4ead736eecb0661384b0272f1d7817d65d2c0728b" Sep 30 09:41:29 crc kubenswrapper[4810]: E0930 09:41:29.840789 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf2578633a5ef11aa2429fe4ead736eecb0661384b0272f1d7817d65d2c0728b\": container with ID starting with cf2578633a5ef11aa2429fe4ead736eecb0661384b0272f1d7817d65d2c0728b not found: ID does not exist" containerID="cf2578633a5ef11aa2429fe4ead736eecb0661384b0272f1d7817d65d2c0728b" Sep 30 09:41:29 crc kubenswrapper[4810]: I0930 09:41:29.840823 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf2578633a5ef11aa2429fe4ead736eecb0661384b0272f1d7817d65d2c0728b"} err="failed to get container status \"cf2578633a5ef11aa2429fe4ead736eecb0661384b0272f1d7817d65d2c0728b\": rpc error: code = NotFound desc = could not find container \"cf2578633a5ef11aa2429fe4ead736eecb0661384b0272f1d7817d65d2c0728b\": container with ID starting with cf2578633a5ef11aa2429fe4ead736eecb0661384b0272f1d7817d65d2c0728b not found: ID does not exist" Sep 30 09:41:30 crc kubenswrapper[4810]: I0930 09:41:30.066305 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n5cvg"] Sep 30 09:41:30 crc kubenswrapper[4810]: I0930 09:41:30.074659 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n5cvg"] Sep 30 09:41:31 crc kubenswrapper[4810]: I0930 09:41:31.323464 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1167049c-b656-4f5e-ba9f-ae8247a60752" path="/var/lib/kubelet/pods/1167049c-b656-4f5e-ba9f-ae8247a60752/volumes" Sep 30 09:41:56 crc kubenswrapper[4810]: I0930 09:41:56.017038 4810 generic.go:334] "Generic (PLEG): container finished" podID="439949d3-6a38-4e79-bded-bee33d6884ef" containerID="d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f" exitCode=0 Sep 30 09:41:56 crc kubenswrapper[4810]: I0930 09:41:56.017158 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48zlz/must-gather-jj5v6" event={"ID":"439949d3-6a38-4e79-bded-bee33d6884ef","Type":"ContainerDied","Data":"d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f"} Sep 30 09:41:56 crc kubenswrapper[4810]: I0930 09:41:56.018819 4810 scope.go:117] "RemoveContainer" containerID="d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f" Sep 30 09:41:56 crc kubenswrapper[4810]: I0930 09:41:56.108157 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-48zlz_must-gather-jj5v6_439949d3-6a38-4e79-bded-bee33d6884ef/gather/0.log" Sep 30 09:42:04 crc kubenswrapper[4810]: I0930 09:42:04.648592 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-48zlz/must-gather-jj5v6"] Sep 30 09:42:04 crc kubenswrapper[4810]: I0930 09:42:04.649946 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-48zlz/must-gather-jj5v6" podUID="439949d3-6a38-4e79-bded-bee33d6884ef" containerName="copy" containerID="cri-o://300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5" gracePeriod=2 Sep 30 09:42:04 crc kubenswrapper[4810]: I0930 09:42:04.652227 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-48zlz/must-gather-jj5v6"] Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.099397 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-48zlz_must-gather-jj5v6_439949d3-6a38-4e79-bded-bee33d6884ef/copy/0.log" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.099958 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/must-gather-jj5v6" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.122373 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-48zlz_must-gather-jj5v6_439949d3-6a38-4e79-bded-bee33d6884ef/copy/0.log" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.122878 4810 generic.go:334] "Generic (PLEG): container finished" podID="439949d3-6a38-4e79-bded-bee33d6884ef" containerID="300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5" exitCode=143 Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.122940 4810 scope.go:117] "RemoveContainer" containerID="300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.122938 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48zlz/must-gather-jj5v6" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.142849 4810 scope.go:117] "RemoveContainer" containerID="d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.164896 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/439949d3-6a38-4e79-bded-bee33d6884ef-must-gather-output\") pod \"439949d3-6a38-4e79-bded-bee33d6884ef\" (UID: \"439949d3-6a38-4e79-bded-bee33d6884ef\") " Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.165050 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-427g2\" (UniqueName: \"kubernetes.io/projected/439949d3-6a38-4e79-bded-bee33d6884ef-kube-api-access-427g2\") pod \"439949d3-6a38-4e79-bded-bee33d6884ef\" (UID: \"439949d3-6a38-4e79-bded-bee33d6884ef\") " Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.172946 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/439949d3-6a38-4e79-bded-bee33d6884ef-kube-api-access-427g2" (OuterVolumeSpecName: "kube-api-access-427g2") pod "439949d3-6a38-4e79-bded-bee33d6884ef" (UID: "439949d3-6a38-4e79-bded-bee33d6884ef"). InnerVolumeSpecName "kube-api-access-427g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.224221 4810 scope.go:117] "RemoveContainer" containerID="300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5" Sep 30 09:42:05 crc kubenswrapper[4810]: E0930 09:42:05.227744 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5\": container with ID starting with 300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5 not found: ID does not exist" containerID="300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.227990 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5"} err="failed to get container status \"300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5\": rpc error: code = NotFound desc = could not find container \"300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5\": container with ID starting with 300aaa4a36bc3c470efa5fe05d1efbccc5efb63111906bb45f79dd879495ffe5 not found: ID does not exist" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.228043 4810 scope.go:117] "RemoveContainer" containerID="d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f" Sep 30 09:42:05 crc kubenswrapper[4810]: E0930 09:42:05.229334 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f\": container with ID starting with d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f not found: ID does not exist" containerID="d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.229386 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f"} err="failed to get container status \"d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f\": rpc error: code = NotFound desc = could not find container \"d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f\": container with ID starting with d68b914b060e3ac675153dfb48e54b1dcbee732c78e2b7ba56db833c10f0763f not found: ID does not exist" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.267646 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-427g2\" (UniqueName: \"kubernetes.io/projected/439949d3-6a38-4e79-bded-bee33d6884ef-kube-api-access-427g2\") on node \"crc\" DevicePath \"\"" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.371916 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/439949d3-6a38-4e79-bded-bee33d6884ef-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "439949d3-6a38-4e79-bded-bee33d6884ef" (UID: "439949d3-6a38-4e79-bded-bee33d6884ef"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:42:05 crc kubenswrapper[4810]: I0930 09:42:05.376337 4810 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/439949d3-6a38-4e79-bded-bee33d6884ef-must-gather-output\") on node \"crc\" DevicePath \"\"" Sep 30 09:42:07 crc kubenswrapper[4810]: I0930 09:42:07.318700 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="439949d3-6a38-4e79-bded-bee33d6884ef" path="/var/lib/kubelet/pods/439949d3-6a38-4e79-bded-bee33d6884ef/volumes" Sep 30 09:42:15 crc kubenswrapper[4810]: I0930 09:42:15.912043 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:42:15 crc kubenswrapper[4810]: I0930 09:42:15.912584 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:42:45 crc kubenswrapper[4810]: I0930 09:42:45.911726 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:42:45 crc kubenswrapper[4810]: I0930 09:42:45.912232 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.722353 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pdntm/must-gather-2vvzs"] Sep 30 09:42:55 crc kubenswrapper[4810]: E0930 09:42:55.723515 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1167049c-b656-4f5e-ba9f-ae8247a60752" containerName="extract-content" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.723536 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1167049c-b656-4f5e-ba9f-ae8247a60752" containerName="extract-content" Sep 30 09:42:55 crc kubenswrapper[4810]: E0930 09:42:55.723563 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="439949d3-6a38-4e79-bded-bee33d6884ef" containerName="copy" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.723571 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="439949d3-6a38-4e79-bded-bee33d6884ef" containerName="copy" Sep 30 09:42:55 crc kubenswrapper[4810]: E0930 09:42:55.723589 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1167049c-b656-4f5e-ba9f-ae8247a60752" containerName="registry-server" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.723597 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1167049c-b656-4f5e-ba9f-ae8247a60752" containerName="registry-server" Sep 30 09:42:55 crc kubenswrapper[4810]: E0930 09:42:55.723630 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="439949d3-6a38-4e79-bded-bee33d6884ef" containerName="gather" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.723637 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="439949d3-6a38-4e79-bded-bee33d6884ef" containerName="gather" Sep 30 09:42:55 crc kubenswrapper[4810]: E0930 09:42:55.723659 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1167049c-b656-4f5e-ba9f-ae8247a60752" containerName="extract-utilities" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.723667 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1167049c-b656-4f5e-ba9f-ae8247a60752" containerName="extract-utilities" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.723917 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="439949d3-6a38-4e79-bded-bee33d6884ef" containerName="gather" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.723938 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1167049c-b656-4f5e-ba9f-ae8247a60752" containerName="registry-server" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.723973 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="439949d3-6a38-4e79-bded-bee33d6884ef" containerName="copy" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.725190 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/must-gather-2vvzs" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.737975 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pdntm"/"openshift-service-ca.crt" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.738168 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pdntm"/"kube-root-ca.crt" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.738626 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pdntm/must-gather-2vvzs"] Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.846053 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mppgh\" (UniqueName: \"kubernetes.io/projected/75f3fd61-26a0-4cb9-bdba-144e953d81e8-kube-api-access-mppgh\") pod \"must-gather-2vvzs\" (UID: \"75f3fd61-26a0-4cb9-bdba-144e953d81e8\") " pod="openshift-must-gather-pdntm/must-gather-2vvzs" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.846151 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/75f3fd61-26a0-4cb9-bdba-144e953d81e8-must-gather-output\") pod \"must-gather-2vvzs\" (UID: \"75f3fd61-26a0-4cb9-bdba-144e953d81e8\") " pod="openshift-must-gather-pdntm/must-gather-2vvzs" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.947774 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mppgh\" (UniqueName: \"kubernetes.io/projected/75f3fd61-26a0-4cb9-bdba-144e953d81e8-kube-api-access-mppgh\") pod \"must-gather-2vvzs\" (UID: \"75f3fd61-26a0-4cb9-bdba-144e953d81e8\") " pod="openshift-must-gather-pdntm/must-gather-2vvzs" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.947846 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/75f3fd61-26a0-4cb9-bdba-144e953d81e8-must-gather-output\") pod \"must-gather-2vvzs\" (UID: \"75f3fd61-26a0-4cb9-bdba-144e953d81e8\") " pod="openshift-must-gather-pdntm/must-gather-2vvzs" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.948369 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/75f3fd61-26a0-4cb9-bdba-144e953d81e8-must-gather-output\") pod \"must-gather-2vvzs\" (UID: \"75f3fd61-26a0-4cb9-bdba-144e953d81e8\") " pod="openshift-must-gather-pdntm/must-gather-2vvzs" Sep 30 09:42:55 crc kubenswrapper[4810]: I0930 09:42:55.964714 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mppgh\" (UniqueName: \"kubernetes.io/projected/75f3fd61-26a0-4cb9-bdba-144e953d81e8-kube-api-access-mppgh\") pod \"must-gather-2vvzs\" (UID: \"75f3fd61-26a0-4cb9-bdba-144e953d81e8\") " pod="openshift-must-gather-pdntm/must-gather-2vvzs" Sep 30 09:42:56 crc kubenswrapper[4810]: I0930 09:42:56.053241 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/must-gather-2vvzs" Sep 30 09:42:56 crc kubenswrapper[4810]: I0930 09:42:56.548938 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pdntm/must-gather-2vvzs"] Sep 30 09:42:56 crc kubenswrapper[4810]: I0930 09:42:56.723909 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/must-gather-2vvzs" event={"ID":"75f3fd61-26a0-4cb9-bdba-144e953d81e8","Type":"ContainerStarted","Data":"1ebb798f7809d0b539f67d491be1a0addbd48918e5c244e27748802fcb59a2b7"} Sep 30 09:42:57 crc kubenswrapper[4810]: I0930 09:42:57.735519 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/must-gather-2vvzs" event={"ID":"75f3fd61-26a0-4cb9-bdba-144e953d81e8","Type":"ContainerStarted","Data":"a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560"} Sep 30 09:42:57 crc kubenswrapper[4810]: I0930 09:42:57.735886 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/must-gather-2vvzs" event={"ID":"75f3fd61-26a0-4cb9-bdba-144e953d81e8","Type":"ContainerStarted","Data":"fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0"} Sep 30 09:42:57 crc kubenswrapper[4810]: I0930 09:42:57.756084 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pdntm/must-gather-2vvzs" podStartSLOduration=2.756055978 podStartE2EDuration="2.756055978s" podCreationTimestamp="2025-09-30 09:42:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 09:42:57.751044653 +0000 UTC m=+6001.203243940" watchObservedRunningTime="2025-09-30 09:42:57.756055978 +0000 UTC m=+6001.208255255" Sep 30 09:43:00 crc kubenswrapper[4810]: I0930 09:43:00.459802 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pdntm/crc-debug-mdwmv"] Sep 30 09:43:00 crc kubenswrapper[4810]: I0930 09:43:00.461862 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-mdwmv" Sep 30 09:43:00 crc kubenswrapper[4810]: I0930 09:43:00.466524 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pdntm"/"default-dockercfg-pllwj" Sep 30 09:43:00 crc kubenswrapper[4810]: I0930 09:43:00.573439 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8p8b\" (UniqueName: \"kubernetes.io/projected/281ea86c-1429-40e5-9e26-2f06551a023c-kube-api-access-b8p8b\") pod \"crc-debug-mdwmv\" (UID: \"281ea86c-1429-40e5-9e26-2f06551a023c\") " pod="openshift-must-gather-pdntm/crc-debug-mdwmv" Sep 30 09:43:00 crc kubenswrapper[4810]: I0930 09:43:00.573908 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/281ea86c-1429-40e5-9e26-2f06551a023c-host\") pod \"crc-debug-mdwmv\" (UID: \"281ea86c-1429-40e5-9e26-2f06551a023c\") " pod="openshift-must-gather-pdntm/crc-debug-mdwmv" Sep 30 09:43:00 crc kubenswrapper[4810]: I0930 09:43:00.676058 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/281ea86c-1429-40e5-9e26-2f06551a023c-host\") pod \"crc-debug-mdwmv\" (UID: \"281ea86c-1429-40e5-9e26-2f06551a023c\") " pod="openshift-must-gather-pdntm/crc-debug-mdwmv" Sep 30 09:43:00 crc kubenswrapper[4810]: I0930 09:43:00.676185 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/281ea86c-1429-40e5-9e26-2f06551a023c-host\") pod \"crc-debug-mdwmv\" (UID: \"281ea86c-1429-40e5-9e26-2f06551a023c\") " pod="openshift-must-gather-pdntm/crc-debug-mdwmv" Sep 30 09:43:00 crc kubenswrapper[4810]: I0930 09:43:00.676200 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8p8b\" (UniqueName: \"kubernetes.io/projected/281ea86c-1429-40e5-9e26-2f06551a023c-kube-api-access-b8p8b\") pod \"crc-debug-mdwmv\" (UID: \"281ea86c-1429-40e5-9e26-2f06551a023c\") " pod="openshift-must-gather-pdntm/crc-debug-mdwmv" Sep 30 09:43:00 crc kubenswrapper[4810]: I0930 09:43:00.696677 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8p8b\" (UniqueName: \"kubernetes.io/projected/281ea86c-1429-40e5-9e26-2f06551a023c-kube-api-access-b8p8b\") pod \"crc-debug-mdwmv\" (UID: \"281ea86c-1429-40e5-9e26-2f06551a023c\") " pod="openshift-must-gather-pdntm/crc-debug-mdwmv" Sep 30 09:43:00 crc kubenswrapper[4810]: I0930 09:43:00.801394 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-mdwmv" Sep 30 09:43:00 crc kubenswrapper[4810]: W0930 09:43:00.858840 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod281ea86c_1429_40e5_9e26_2f06551a023c.slice/crio-50cd66f8ed8f692c534af9aa555be051d87b9137fa185098d814c583804a9dea WatchSource:0}: Error finding container 50cd66f8ed8f692c534af9aa555be051d87b9137fa185098d814c583804a9dea: Status 404 returned error can't find the container with id 50cd66f8ed8f692c534af9aa555be051d87b9137fa185098d814c583804a9dea Sep 30 09:43:01 crc kubenswrapper[4810]: I0930 09:43:01.782322 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/crc-debug-mdwmv" event={"ID":"281ea86c-1429-40e5-9e26-2f06551a023c","Type":"ContainerStarted","Data":"57622db53c54a7692bfe9230fe7b91e7386251715175014a2bff2e21c3fa02bc"} Sep 30 09:43:01 crc kubenswrapper[4810]: I0930 09:43:01.783131 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/crc-debug-mdwmv" event={"ID":"281ea86c-1429-40e5-9e26-2f06551a023c","Type":"ContainerStarted","Data":"50cd66f8ed8f692c534af9aa555be051d87b9137fa185098d814c583804a9dea"} Sep 30 09:43:01 crc kubenswrapper[4810]: I0930 09:43:01.803923 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pdntm/crc-debug-mdwmv" podStartSLOduration=1.803897928 podStartE2EDuration="1.803897928s" podCreationTimestamp="2025-09-30 09:43:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 09:43:01.799059678 +0000 UTC m=+6005.251258945" watchObservedRunningTime="2025-09-30 09:43:01.803897928 +0000 UTC m=+6005.256097195" Sep 30 09:43:15 crc kubenswrapper[4810]: I0930 09:43:15.911180 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:43:15 crc kubenswrapper[4810]: I0930 09:43:15.911743 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:43:15 crc kubenswrapper[4810]: I0930 09:43:15.911787 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 09:43:15 crc kubenswrapper[4810]: I0930 09:43:15.912608 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 09:43:15 crc kubenswrapper[4810]: I0930 09:43:15.912667 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" gracePeriod=600 Sep 30 09:43:16 crc kubenswrapper[4810]: E0930 09:43:16.038575 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:43:16 crc kubenswrapper[4810]: I0930 09:43:16.931291 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" exitCode=0 Sep 30 09:43:16 crc kubenswrapper[4810]: I0930 09:43:16.931330 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e"} Sep 30 09:43:16 crc kubenswrapper[4810]: I0930 09:43:16.931363 4810 scope.go:117] "RemoveContainer" containerID="3cfe1e0b8905c291915720129de0e23bdd06ea6549e1c8ca139a56a2878a94dd" Sep 30 09:43:16 crc kubenswrapper[4810]: I0930 09:43:16.932323 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:43:16 crc kubenswrapper[4810]: E0930 09:43:16.932781 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:43:29 crc kubenswrapper[4810]: I0930 09:43:29.307179 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:43:29 crc kubenswrapper[4810]: E0930 09:43:29.307951 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:43:40 crc kubenswrapper[4810]: I0930 09:43:40.306722 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:43:40 crc kubenswrapper[4810]: E0930 09:43:40.307654 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:43:42 crc kubenswrapper[4810]: I0930 09:43:42.777011 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hq8cc"] Sep 30 09:43:42 crc kubenswrapper[4810]: I0930 09:43:42.779373 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:42 crc kubenswrapper[4810]: I0930 09:43:42.801727 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hq8cc"] Sep 30 09:43:42 crc kubenswrapper[4810]: I0930 09:43:42.906646 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-utilities\") pod \"community-operators-hq8cc\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:42 crc kubenswrapper[4810]: I0930 09:43:42.906886 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqm6n\" (UniqueName: \"kubernetes.io/projected/445ea063-bb16-4d55-be71-761b8a7181f6-kube-api-access-qqm6n\") pod \"community-operators-hq8cc\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:42 crc kubenswrapper[4810]: I0930 09:43:42.906949 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-catalog-content\") pod \"community-operators-hq8cc\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:43 crc kubenswrapper[4810]: I0930 09:43:43.009329 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqm6n\" (UniqueName: \"kubernetes.io/projected/445ea063-bb16-4d55-be71-761b8a7181f6-kube-api-access-qqm6n\") pod \"community-operators-hq8cc\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:43 crc kubenswrapper[4810]: I0930 09:43:43.009407 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-catalog-content\") pod \"community-operators-hq8cc\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:43 crc kubenswrapper[4810]: I0930 09:43:43.009586 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-utilities\") pod \"community-operators-hq8cc\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:43 crc kubenswrapper[4810]: I0930 09:43:43.010045 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-catalog-content\") pod \"community-operators-hq8cc\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:43 crc kubenswrapper[4810]: I0930 09:43:43.010097 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-utilities\") pod \"community-operators-hq8cc\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:43 crc kubenswrapper[4810]: I0930 09:43:43.038796 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqm6n\" (UniqueName: \"kubernetes.io/projected/445ea063-bb16-4d55-be71-761b8a7181f6-kube-api-access-qqm6n\") pod \"community-operators-hq8cc\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:43 crc kubenswrapper[4810]: I0930 09:43:43.115084 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:43 crc kubenswrapper[4810]: I0930 09:43:43.776477 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hq8cc"] Sep 30 09:43:44 crc kubenswrapper[4810]: I0930 09:43:44.247979 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hq8cc" event={"ID":"445ea063-bb16-4d55-be71-761b8a7181f6","Type":"ContainerStarted","Data":"fb8e5cb0f1b7a3b827b2ee0459e0982598d10728bde19be45225cfe7e440438c"} Sep 30 09:43:45 crc kubenswrapper[4810]: I0930 09:43:45.258728 4810 generic.go:334] "Generic (PLEG): container finished" podID="445ea063-bb16-4d55-be71-761b8a7181f6" containerID="1f4ffdebc4b3ae1750fc15c3f24cfc099cc3fbeca342240fa45a09afaf40a791" exitCode=0 Sep 30 09:43:45 crc kubenswrapper[4810]: I0930 09:43:45.258929 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hq8cc" event={"ID":"445ea063-bb16-4d55-be71-761b8a7181f6","Type":"ContainerDied","Data":"1f4ffdebc4b3ae1750fc15c3f24cfc099cc3fbeca342240fa45a09afaf40a791"} Sep 30 09:43:47 crc kubenswrapper[4810]: I0930 09:43:47.281188 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hq8cc" event={"ID":"445ea063-bb16-4d55-be71-761b8a7181f6","Type":"ContainerStarted","Data":"6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df"} Sep 30 09:43:49 crc kubenswrapper[4810]: I0930 09:43:49.307429 4810 generic.go:334] "Generic (PLEG): container finished" podID="445ea063-bb16-4d55-be71-761b8a7181f6" containerID="6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df" exitCode=0 Sep 30 09:43:49 crc kubenswrapper[4810]: I0930 09:43:49.318972 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hq8cc" event={"ID":"445ea063-bb16-4d55-be71-761b8a7181f6","Type":"ContainerDied","Data":"6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df"} Sep 30 09:43:50 crc kubenswrapper[4810]: I0930 09:43:50.322250 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hq8cc" event={"ID":"445ea063-bb16-4d55-be71-761b8a7181f6","Type":"ContainerStarted","Data":"f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b"} Sep 30 09:43:50 crc kubenswrapper[4810]: I0930 09:43:50.350198 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hq8cc" podStartSLOduration=3.691532951 podStartE2EDuration="8.350174057s" podCreationTimestamp="2025-09-30 09:43:42 +0000 UTC" firstStartedPulling="2025-09-30 09:43:45.260646914 +0000 UTC m=+6048.712846181" lastFinishedPulling="2025-09-30 09:43:49.91928802 +0000 UTC m=+6053.371487287" observedRunningTime="2025-09-30 09:43:50.341941229 +0000 UTC m=+6053.794140496" watchObservedRunningTime="2025-09-30 09:43:50.350174057 +0000 UTC m=+6053.802373324" Sep 30 09:43:52 crc kubenswrapper[4810]: I0930 09:43:52.307389 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:43:52 crc kubenswrapper[4810]: E0930 09:43:52.307967 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:43:53 crc kubenswrapper[4810]: I0930 09:43:53.116866 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:53 crc kubenswrapper[4810]: I0930 09:43:53.117149 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:43:53 crc kubenswrapper[4810]: I0930 09:43:53.171107 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:44:03 crc kubenswrapper[4810]: I0930 09:44:03.192406 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:44:03 crc kubenswrapper[4810]: I0930 09:44:03.250618 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hq8cc"] Sep 30 09:44:03 crc kubenswrapper[4810]: I0930 09:44:03.484143 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hq8cc" podUID="445ea063-bb16-4d55-be71-761b8a7181f6" containerName="registry-server" containerID="cri-o://f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b" gracePeriod=2 Sep 30 09:44:03 crc kubenswrapper[4810]: I0930 09:44:03.940142 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.095013 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqm6n\" (UniqueName: \"kubernetes.io/projected/445ea063-bb16-4d55-be71-761b8a7181f6-kube-api-access-qqm6n\") pod \"445ea063-bb16-4d55-be71-761b8a7181f6\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.095062 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-utilities\") pod \"445ea063-bb16-4d55-be71-761b8a7181f6\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.095242 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-catalog-content\") pod \"445ea063-bb16-4d55-be71-761b8a7181f6\" (UID: \"445ea063-bb16-4d55-be71-761b8a7181f6\") " Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.096627 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-utilities" (OuterVolumeSpecName: "utilities") pod "445ea063-bb16-4d55-be71-761b8a7181f6" (UID: "445ea063-bb16-4d55-be71-761b8a7181f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.103333 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/445ea063-bb16-4d55-be71-761b8a7181f6-kube-api-access-qqm6n" (OuterVolumeSpecName: "kube-api-access-qqm6n") pod "445ea063-bb16-4d55-be71-761b8a7181f6" (UID: "445ea063-bb16-4d55-be71-761b8a7181f6"). InnerVolumeSpecName "kube-api-access-qqm6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.167592 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "445ea063-bb16-4d55-be71-761b8a7181f6" (UID: "445ea063-bb16-4d55-be71-761b8a7181f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.197754 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqm6n\" (UniqueName: \"kubernetes.io/projected/445ea063-bb16-4d55-be71-761b8a7181f6-kube-api-access-qqm6n\") on node \"crc\" DevicePath \"\"" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.197808 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.197821 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445ea063-bb16-4d55-be71-761b8a7181f6-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.495971 4810 generic.go:334] "Generic (PLEG): container finished" podID="445ea063-bb16-4d55-be71-761b8a7181f6" containerID="f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b" exitCode=0 Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.496055 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hq8cc" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.496052 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hq8cc" event={"ID":"445ea063-bb16-4d55-be71-761b8a7181f6","Type":"ContainerDied","Data":"f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b"} Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.496445 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hq8cc" event={"ID":"445ea063-bb16-4d55-be71-761b8a7181f6","Type":"ContainerDied","Data":"fb8e5cb0f1b7a3b827b2ee0459e0982598d10728bde19be45225cfe7e440438c"} Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.496467 4810 scope.go:117] "RemoveContainer" containerID="f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.543508 4810 scope.go:117] "RemoveContainer" containerID="6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.548545 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hq8cc"] Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.559112 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hq8cc"] Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.587505 4810 scope.go:117] "RemoveContainer" containerID="1f4ffdebc4b3ae1750fc15c3f24cfc099cc3fbeca342240fa45a09afaf40a791" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.646455 4810 scope.go:117] "RemoveContainer" containerID="f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b" Sep 30 09:44:04 crc kubenswrapper[4810]: E0930 09:44:04.649763 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b\": container with ID starting with f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b not found: ID does not exist" containerID="f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.649818 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b"} err="failed to get container status \"f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b\": rpc error: code = NotFound desc = could not find container \"f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b\": container with ID starting with f47b5b3d90eb4282d39b5bb19cc023475ea67cfa7e8e40ed9c37346248b1616b not found: ID does not exist" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.649860 4810 scope.go:117] "RemoveContainer" containerID="6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df" Sep 30 09:44:04 crc kubenswrapper[4810]: E0930 09:44:04.650442 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df\": container with ID starting with 6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df not found: ID does not exist" containerID="6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.650471 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df"} err="failed to get container status \"6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df\": rpc error: code = NotFound desc = could not find container \"6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df\": container with ID starting with 6ffeced47a2c377f2087ae30ef9b6c75312b5c29b99cc2f7ae874fcc69a052df not found: ID does not exist" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.650491 4810 scope.go:117] "RemoveContainer" containerID="1f4ffdebc4b3ae1750fc15c3f24cfc099cc3fbeca342240fa45a09afaf40a791" Sep 30 09:44:04 crc kubenswrapper[4810]: E0930 09:44:04.650827 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f4ffdebc4b3ae1750fc15c3f24cfc099cc3fbeca342240fa45a09afaf40a791\": container with ID starting with 1f4ffdebc4b3ae1750fc15c3f24cfc099cc3fbeca342240fa45a09afaf40a791 not found: ID does not exist" containerID="1f4ffdebc4b3ae1750fc15c3f24cfc099cc3fbeca342240fa45a09afaf40a791" Sep 30 09:44:04 crc kubenswrapper[4810]: I0930 09:44:04.650854 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f4ffdebc4b3ae1750fc15c3f24cfc099cc3fbeca342240fa45a09afaf40a791"} err="failed to get container status \"1f4ffdebc4b3ae1750fc15c3f24cfc099cc3fbeca342240fa45a09afaf40a791\": rpc error: code = NotFound desc = could not find container \"1f4ffdebc4b3ae1750fc15c3f24cfc099cc3fbeca342240fa45a09afaf40a791\": container with ID starting with 1f4ffdebc4b3ae1750fc15c3f24cfc099cc3fbeca342240fa45a09afaf40a791 not found: ID does not exist" Sep 30 09:44:05 crc kubenswrapper[4810]: I0930 09:44:05.307178 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:44:05 crc kubenswrapper[4810]: E0930 09:44:05.307479 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:44:05 crc kubenswrapper[4810]: I0930 09:44:05.317389 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="445ea063-bb16-4d55-be71-761b8a7181f6" path="/var/lib/kubelet/pods/445ea063-bb16-4d55-be71-761b8a7181f6/volumes" Sep 30 09:44:17 crc kubenswrapper[4810]: I0930 09:44:17.345998 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:44:17 crc kubenswrapper[4810]: E0930 09:44:17.346891 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:44:22 crc kubenswrapper[4810]: I0930 09:44:22.080597 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c6d89f69d-k9g44_23454188-6622-4e7c-90f0-2fddf031738e/barbican-api/0.log" Sep 30 09:44:22 crc kubenswrapper[4810]: I0930 09:44:22.148589 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c6d89f69d-k9g44_23454188-6622-4e7c-90f0-2fddf031738e/barbican-api-log/0.log" Sep 30 09:44:22 crc kubenswrapper[4810]: I0930 09:44:22.314021 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5dff5594b-jj7gs_d2c5a1aa-c611-4702-a0bb-27693ebb7c7a/barbican-keystone-listener/0.log" Sep 30 09:44:22 crc kubenswrapper[4810]: I0930 09:44:22.420007 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5dff5594b-jj7gs_d2c5a1aa-c611-4702-a0bb-27693ebb7c7a/barbican-keystone-listener-log/0.log" Sep 30 09:44:22 crc kubenswrapper[4810]: I0930 09:44:22.539601 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-687444b9dc-hxqlq_ff274965-178c-4eff-9841-494634e711f0/barbican-worker/0.log" Sep 30 09:44:22 crc kubenswrapper[4810]: I0930 09:44:22.643655 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-687444b9dc-hxqlq_ff274965-178c-4eff-9841-494634e711f0/barbican-worker-log/0.log" Sep 30 09:44:22 crc kubenswrapper[4810]: I0930 09:44:22.762119 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-57ssz_6465dd71-b2e8-4a11-90a8-406ddd763a0c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:23 crc kubenswrapper[4810]: I0930 09:44:23.053207 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ae076b6e-09cd-4911-89c3-e6edb61516ae/ceilometer-central-agent/0.log" Sep 30 09:44:23 crc kubenswrapper[4810]: I0930 09:44:23.058924 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ae076b6e-09cd-4911-89c3-e6edb61516ae/ceilometer-notification-agent/0.log" Sep 30 09:44:23 crc kubenswrapper[4810]: I0930 09:44:23.093996 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ae076b6e-09cd-4911-89c3-e6edb61516ae/proxy-httpd/0.log" Sep 30 09:44:23 crc kubenswrapper[4810]: I0930 09:44:23.244135 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ae076b6e-09cd-4911-89c3-e6edb61516ae/sg-core/0.log" Sep 30 09:44:23 crc kubenswrapper[4810]: I0930 09:44:23.441257 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_58537813-eea7-4a58-a900-f3473797d7d2/cinder-api-log/0.log" Sep 30 09:44:23 crc kubenswrapper[4810]: I0930 09:44:23.616841 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_58537813-eea7-4a58-a900-f3473797d7d2/cinder-api/0.log" Sep 30 09:44:23 crc kubenswrapper[4810]: I0930 09:44:23.691994 4810 scope.go:117] "RemoveContainer" containerID="88a0559833e2aadd014203d1bdfe41327bbf9ad4a049400611aa1a291ed42f96" Sep 30 09:44:23 crc kubenswrapper[4810]: I0930 09:44:23.693494 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e10f507-33b5-4477-882f-6c0b3271034d/cinder-scheduler/0.log" Sep 30 09:44:23 crc kubenswrapper[4810]: I0930 09:44:23.841117 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e10f507-33b5-4477-882f-6c0b3271034d/probe/0.log" Sep 30 09:44:23 crc kubenswrapper[4810]: I0930 09:44:23.932753 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-22dr4_459ff4e4-7043-4e6e-9774-97dab472459c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:24 crc kubenswrapper[4810]: I0930 09:44:24.063105 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hhtnx_80653750-cd3d-4e95-b4a1-4909c325c34c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:24 crc kubenswrapper[4810]: I0930 09:44:24.260911 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-585f78d44c-nlbbz_cf951386-0eef-4a7f-9247-b854b9a39d7b/init/0.log" Sep 30 09:44:24 crc kubenswrapper[4810]: I0930 09:44:24.476925 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-585f78d44c-nlbbz_cf951386-0eef-4a7f-9247-b854b9a39d7b/init/0.log" Sep 30 09:44:24 crc kubenswrapper[4810]: I0930 09:44:24.641292 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-585f78d44c-nlbbz_cf951386-0eef-4a7f-9247-b854b9a39d7b/dnsmasq-dns/0.log" Sep 30 09:44:24 crc kubenswrapper[4810]: I0930 09:44:24.690362 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-7wfjq_3f13a98d-9385-4e17-a8ef-fd0dec1866ac/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:24 crc kubenswrapper[4810]: I0930 09:44:24.845549 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f35de952-dcbd-422d-bbdb-b8995de342a1/glance-httpd/0.log" Sep 30 09:44:24 crc kubenswrapper[4810]: I0930 09:44:24.858720 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f35de952-dcbd-422d-bbdb-b8995de342a1/glance-log/0.log" Sep 30 09:44:25 crc kubenswrapper[4810]: I0930 09:44:25.057000 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_77634c10-6e69-4f6c-9090-be8a59588745/glance-log/0.log" Sep 30 09:44:25 crc kubenswrapper[4810]: I0930 09:44:25.059281 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_77634c10-6e69-4f6c-9090-be8a59588745/glance-httpd/0.log" Sep 30 09:44:25 crc kubenswrapper[4810]: I0930 09:44:25.353409 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d769bd5f4-zkxj4_a71102a5-388b-4dcb-bd80-db92ed28f1f5/horizon/0.log" Sep 30 09:44:25 crc kubenswrapper[4810]: I0930 09:44:25.562866 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-2rltf_d2e22e53-6826-4a58-abb0-eb2d279d2b1c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:25 crc kubenswrapper[4810]: I0930 09:44:25.667122 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-9gnn7_75b17470-760f-4fa6-9550-c2281ed5ae6b/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:25 crc kubenswrapper[4810]: I0930 09:44:25.805117 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29320381-wgv24_dd37005c-311f-4aef-9f91-bc01d5d62249/keystone-cron/0.log" Sep 30 09:44:25 crc kubenswrapper[4810]: I0930 09:44:25.950595 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d769bd5f4-zkxj4_a71102a5-388b-4dcb-bd80-db92ed28f1f5/horizon-log/0.log" Sep 30 09:44:26 crc kubenswrapper[4810]: I0930 09:44:26.144028 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b0778b82-d141-47f3-b766-a3df63f09d12/kube-state-metrics/0.log" Sep 30 09:44:26 crc kubenswrapper[4810]: I0930 09:44:26.383837 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-xzwzd_9b0be1a4-a8dc-4b69-ab6f-ce080caec2a1/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:26 crc kubenswrapper[4810]: I0930 09:44:26.709878 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-f95cc94f6-tz5ks_2d05e909-d2bc-49d7-af7b-0736de4c3f95/keystone-api/0.log" Sep 30 09:44:27 crc kubenswrapper[4810]: I0930 09:44:27.033848 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f769c46d7-mgh2q_34fe741b-5b18-4fe8-9f54-1fd3e62d9e24/neutron-api/0.log" Sep 30 09:44:27 crc kubenswrapper[4810]: I0930 09:44:27.134844 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f769c46d7-mgh2q_34fe741b-5b18-4fe8-9f54-1fd3e62d9e24/neutron-httpd/0.log" Sep 30 09:44:27 crc kubenswrapper[4810]: I0930 09:44:27.154354 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-fdt9z_5de6b629-1489-4099-89bf-7773b0bdaeb8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:28 crc kubenswrapper[4810]: I0930 09:44:28.066628 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_8f67f47e-4e54-44b8-9596-c612858accdd/nova-cell0-conductor-conductor/0.log" Sep 30 09:44:28 crc kubenswrapper[4810]: I0930 09:44:28.720022 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_4342c2f0-9d27-4211-a302-a1822255f011/nova-cell1-conductor-conductor/0.log" Sep 30 09:44:28 crc kubenswrapper[4810]: I0930 09:44:28.886940 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_7e2cfe82-5249-4207-9178-f4ebb56a9f93/nova-api-log/0.log" Sep 30 09:44:29 crc kubenswrapper[4810]: I0930 09:44:29.284827 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_7e2cfe82-5249-4207-9178-f4ebb56a9f93/nova-api-api/0.log" Sep 30 09:44:29 crc kubenswrapper[4810]: I0930 09:44:29.333831 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_fc321aeb-240d-414e-91de-ba95671d6556/nova-cell1-novncproxy-novncproxy/0.log" Sep 30 09:44:29 crc kubenswrapper[4810]: I0930 09:44:29.482990 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-v6q68_8a25c107-d531-4a79-97e9-7201d6a8485a/nova-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:29 crc kubenswrapper[4810]: I0930 09:44:29.739724 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_3d733771-3677-4a2e-9eda-86144efccf8d/nova-metadata-log/0.log" Sep 30 09:44:30 crc kubenswrapper[4810]: I0930 09:44:30.194163 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cb4aa97-f850-4e7f-a1e1-046bd5235109/mysql-bootstrap/0.log" Sep 30 09:44:30 crc kubenswrapper[4810]: I0930 09:44:30.282791 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_97a5417c-b645-430c-ada6-ae56ca4aa813/nova-scheduler-scheduler/0.log" Sep 30 09:44:30 crc kubenswrapper[4810]: I0930 09:44:30.357973 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cb4aa97-f850-4e7f-a1e1-046bd5235109/mysql-bootstrap/0.log" Sep 30 09:44:30 crc kubenswrapper[4810]: I0930 09:44:30.523371 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cb4aa97-f850-4e7f-a1e1-046bd5235109/galera/0.log" Sep 30 09:44:30 crc kubenswrapper[4810]: I0930 09:44:30.759253 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3ca490c3-ff2f-4553-9e66-4fb456d3a1f2/mysql-bootstrap/0.log" Sep 30 09:44:30 crc kubenswrapper[4810]: I0930 09:44:30.929701 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3ca490c3-ff2f-4553-9e66-4fb456d3a1f2/mysql-bootstrap/0.log" Sep 30 09:44:30 crc kubenswrapper[4810]: I0930 09:44:30.990740 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3ca490c3-ff2f-4553-9e66-4fb456d3a1f2/galera/0.log" Sep 30 09:44:31 crc kubenswrapper[4810]: I0930 09:44:31.217777 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f0aa786d-3e79-423b-82d1-5e9c70ca7c72/openstackclient/0.log" Sep 30 09:44:31 crc kubenswrapper[4810]: I0930 09:44:31.628111 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-9q2d8_af34271b-1a44-4240-88ee-b86490744353/ovn-controller/0.log" Sep 30 09:44:31 crc kubenswrapper[4810]: I0930 09:44:31.829061 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-6wrgl_4b9a3337-b2c2-49c8-9755-81660bb506af/openstack-network-exporter/0.log" Sep 30 09:44:32 crc kubenswrapper[4810]: I0930 09:44:32.056725 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mt698_384b3d69-0ca1-4729-86f7-6d4455fb7ccc/ovsdb-server-init/0.log" Sep 30 09:44:32 crc kubenswrapper[4810]: I0930 09:44:32.260460 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mt698_384b3d69-0ca1-4729-86f7-6d4455fb7ccc/ovsdb-server-init/0.log" Sep 30 09:44:32 crc kubenswrapper[4810]: I0930 09:44:32.306164 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:44:32 crc kubenswrapper[4810]: E0930 09:44:32.306491 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:44:32 crc kubenswrapper[4810]: I0930 09:44:32.434582 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mt698_384b3d69-0ca1-4729-86f7-6d4455fb7ccc/ovsdb-server/0.log" Sep 30 09:44:32 crc kubenswrapper[4810]: I0930 09:44:32.453795 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_3d733771-3677-4a2e-9eda-86144efccf8d/nova-metadata-metadata/0.log" Sep 30 09:44:32 crc kubenswrapper[4810]: I0930 09:44:32.611698 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mt698_384b3d69-0ca1-4729-86f7-6d4455fb7ccc/ovs-vswitchd/0.log" Sep 30 09:44:32 crc kubenswrapper[4810]: I0930 09:44:32.725857 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-5xm2r_dbee734f-4300-41cc-9bb9-fa005daa1df0/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:32 crc kubenswrapper[4810]: I0930 09:44:32.901019 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_306e4bf0-015f-4c7a-ba49-efda86e72442/openstack-network-exporter/0.log" Sep 30 09:44:32 crc kubenswrapper[4810]: I0930 09:44:32.967681 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_306e4bf0-015f-4c7a-ba49-efda86e72442/ovn-northd/0.log" Sep 30 09:44:33 crc kubenswrapper[4810]: I0930 09:44:33.087347 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c94c8fec-2762-469d-86f9-a4aca5c242e5/openstack-network-exporter/0.log" Sep 30 09:44:33 crc kubenswrapper[4810]: I0930 09:44:33.197852 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c94c8fec-2762-469d-86f9-a4aca5c242e5/ovsdbserver-nb/0.log" Sep 30 09:44:33 crc kubenswrapper[4810]: I0930 09:44:33.338151 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ba947eb7-f01f-40ef-b276-b5b73fee9e0a/openstack-network-exporter/0.log" Sep 30 09:44:33 crc kubenswrapper[4810]: I0930 09:44:33.461596 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ba947eb7-f01f-40ef-b276-b5b73fee9e0a/ovsdbserver-sb/0.log" Sep 30 09:44:33 crc kubenswrapper[4810]: I0930 09:44:33.719243 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-54b65d888b-2hkpr_35128b6f-4e9c-414b-847f-773c1418f5e2/placement-api/0.log" Sep 30 09:44:33 crc kubenswrapper[4810]: I0930 09:44:33.956565 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-54b65d888b-2hkpr_35128b6f-4e9c-414b-847f-773c1418f5e2/placement-log/0.log" Sep 30 09:44:33 crc kubenswrapper[4810]: I0930 09:44:33.977888 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ad43140a-2b1c-443b-9595-901c71e14f0d/init-config-reloader/0.log" Sep 30 09:44:34 crc kubenswrapper[4810]: I0930 09:44:34.182818 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ad43140a-2b1c-443b-9595-901c71e14f0d/init-config-reloader/0.log" Sep 30 09:44:34 crc kubenswrapper[4810]: I0930 09:44:34.190657 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ad43140a-2b1c-443b-9595-901c71e14f0d/config-reloader/0.log" Sep 30 09:44:34 crc kubenswrapper[4810]: I0930 09:44:34.251347 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ad43140a-2b1c-443b-9595-901c71e14f0d/prometheus/0.log" Sep 30 09:44:34 crc kubenswrapper[4810]: I0930 09:44:34.355485 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ad43140a-2b1c-443b-9595-901c71e14f0d/thanos-sidecar/0.log" Sep 30 09:44:34 crc kubenswrapper[4810]: I0930 09:44:34.465941 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0344203-4b62-409c-af24-0e619dfad39a/setup-container/0.log" Sep 30 09:44:34 crc kubenswrapper[4810]: I0930 09:44:34.720920 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0344203-4b62-409c-af24-0e619dfad39a/setup-container/0.log" Sep 30 09:44:34 crc kubenswrapper[4810]: I0930 09:44:34.777946 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0344203-4b62-409c-af24-0e619dfad39a/rabbitmq/0.log" Sep 30 09:44:34 crc kubenswrapper[4810]: I0930 09:44:34.919811 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_fcf1eb6a-de6d-48ac-bbf5-4e015c452e18/setup-container/0.log" Sep 30 09:44:35 crc kubenswrapper[4810]: I0930 09:44:35.136090 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_fcf1eb6a-de6d-48ac-bbf5-4e015c452e18/setup-container/0.log" Sep 30 09:44:35 crc kubenswrapper[4810]: I0930 09:44:35.187644 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_fcf1eb6a-de6d-48ac-bbf5-4e015c452e18/rabbitmq/0.log" Sep 30 09:44:35 crc kubenswrapper[4810]: I0930 09:44:35.347159 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b695439c-7bde-408e-b90a-0aa0ff0bc494/setup-container/0.log" Sep 30 09:44:35 crc kubenswrapper[4810]: I0930 09:44:35.559409 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b695439c-7bde-408e-b90a-0aa0ff0bc494/rabbitmq/0.log" Sep 30 09:44:35 crc kubenswrapper[4810]: I0930 09:44:35.637956 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b695439c-7bde-408e-b90a-0aa0ff0bc494/setup-container/0.log" Sep 30 09:44:35 crc kubenswrapper[4810]: I0930 09:44:35.815005 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-fxvbp_6da1c758-a003-49dd-95de-f47d13e11c0a/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:35 crc kubenswrapper[4810]: I0930 09:44:35.862908 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-vhxfz_dc05f1d1-59b0-45d8-8a52-e0cc456a1207/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:36 crc kubenswrapper[4810]: I0930 09:44:36.139537 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-nrg4r_93d196dc-dc56-4116-bb59-d76601947afd/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:36 crc kubenswrapper[4810]: I0930 09:44:36.326050 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-v8zdn_48233a79-011d-4b63-8f0c-885c651ddbee/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:36 crc kubenswrapper[4810]: I0930 09:44:36.441197 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-6gw9x_d5b6b2ff-6366-48ba-9328-7d99a9da7e08/ssh-known-hosts-edpm-deployment/0.log" Sep 30 09:44:36 crc kubenswrapper[4810]: I0930 09:44:36.724850 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5cd556568c-9c6km_dc1aa89c-399a-4db2-a0b1-df1185b14c48/proxy-server/0.log" Sep 30 09:44:36 crc kubenswrapper[4810]: I0930 09:44:36.912295 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5cd556568c-9c6km_dc1aa89c-399a-4db2-a0b1-df1185b14c48/proxy-httpd/0.log" Sep 30 09:44:36 crc kubenswrapper[4810]: I0930 09:44:36.949902 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-zr8dp_cf76de28-c7b8-474f-93a7-5af536e817e8/swift-ring-rebalance/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.080053 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/account-auditor/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.237216 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/account-reaper/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.277595 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/account-replicator/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.293941 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/account-server/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.452278 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/container-auditor/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.464052 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/container-server/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.509483 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/container-replicator/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.674183 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/object-auditor/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.687699 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/container-updater/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.738638 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/object-expirer/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.898671 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/object-replicator/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.939447 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/object-server/0.log" Sep 30 09:44:37 crc kubenswrapper[4810]: I0930 09:44:37.958208 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/object-updater/0.log" Sep 30 09:44:38 crc kubenswrapper[4810]: I0930 09:44:38.098805 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/rsync/0.log" Sep 30 09:44:38 crc kubenswrapper[4810]: I0930 09:44:38.199123 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_edd67220-e97a-465e-9671-2cfcf760475a/swift-recon-cron/0.log" Sep 30 09:44:38 crc kubenswrapper[4810]: I0930 09:44:38.390614 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-kpr4r_ba3931a9-05a1-47f8-a7f7-f41a477a164b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:38 crc kubenswrapper[4810]: I0930 09:44:38.451719 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_78242ee7-a6aa-4087-832e-4834a7f4751f/tempest-tests-tempest-tests-runner/0.log" Sep 30 09:44:38 crc kubenswrapper[4810]: I0930 09:44:38.608653 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_1d54c982-ce8e-43cb-ab4f-69a3cbaebdda/test-operator-logs-container/0.log" Sep 30 09:44:38 crc kubenswrapper[4810]: I0930 09:44:38.877470 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-cdg8b_a86849c2-c174-4b69-84a0-281bb1afda6d/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 09:44:39 crc kubenswrapper[4810]: I0930 09:44:39.995112 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_d04899be-cfe2-4870-ad3c-735148e7bcdd/watcher-applier/0.log" Sep 30 09:44:40 crc kubenswrapper[4810]: I0930 09:44:40.236746 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_32ede37e-d419-4c76-932e-af1de102a9ad/watcher-api-log/0.log" Sep 30 09:44:43 crc kubenswrapper[4810]: I0930 09:44:43.306091 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:44:43 crc kubenswrapper[4810]: E0930 09:44:43.306752 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:44:43 crc kubenswrapper[4810]: I0930 09:44:43.708314 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_5c9ec55c-ddec-4cab-8041-6b5a9c192212/watcher-decision-engine/0.log" Sep 30 09:44:44 crc kubenswrapper[4810]: I0930 09:44:44.410598 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_32ede37e-d419-4c76-932e-af1de102a9ad/watcher-api/0.log" Sep 30 09:44:55 crc kubenswrapper[4810]: I0930 09:44:55.981637 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_f947e702-a475-413a-8e34-08702649bdae/memcached/0.log" Sep 30 09:44:57 crc kubenswrapper[4810]: I0930 09:44:57.314032 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:44:57 crc kubenswrapper[4810]: E0930 09:44:57.315856 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.165104 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq"] Sep 30 09:45:00 crc kubenswrapper[4810]: E0930 09:45:00.166726 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445ea063-bb16-4d55-be71-761b8a7181f6" containerName="registry-server" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.166747 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="445ea063-bb16-4d55-be71-761b8a7181f6" containerName="registry-server" Sep 30 09:45:00 crc kubenswrapper[4810]: E0930 09:45:00.166766 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445ea063-bb16-4d55-be71-761b8a7181f6" containerName="extract-utilities" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.166774 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="445ea063-bb16-4d55-be71-761b8a7181f6" containerName="extract-utilities" Sep 30 09:45:00 crc kubenswrapper[4810]: E0930 09:45:00.166790 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445ea063-bb16-4d55-be71-761b8a7181f6" containerName="extract-content" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.166797 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="445ea063-bb16-4d55-be71-761b8a7181f6" containerName="extract-content" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.167097 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="445ea063-bb16-4d55-be71-761b8a7181f6" containerName="registry-server" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.168049 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.173054 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.173325 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.179283 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq"] Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.224668 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6defdc7-3e4e-497d-933c-896f079bc208-secret-volume\") pod \"collect-profiles-29320425-bm8wq\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.224865 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8fpt\" (UniqueName: \"kubernetes.io/projected/e6defdc7-3e4e-497d-933c-896f079bc208-kube-api-access-c8fpt\") pod \"collect-profiles-29320425-bm8wq\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.224916 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6defdc7-3e4e-497d-933c-896f079bc208-config-volume\") pod \"collect-profiles-29320425-bm8wq\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.327127 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8fpt\" (UniqueName: \"kubernetes.io/projected/e6defdc7-3e4e-497d-933c-896f079bc208-kube-api-access-c8fpt\") pod \"collect-profiles-29320425-bm8wq\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.327179 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6defdc7-3e4e-497d-933c-896f079bc208-config-volume\") pod \"collect-profiles-29320425-bm8wq\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.327299 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6defdc7-3e4e-497d-933c-896f079bc208-secret-volume\") pod \"collect-profiles-29320425-bm8wq\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.328197 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6defdc7-3e4e-497d-933c-896f079bc208-config-volume\") pod \"collect-profiles-29320425-bm8wq\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.338958 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6defdc7-3e4e-497d-933c-896f079bc208-secret-volume\") pod \"collect-profiles-29320425-bm8wq\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.342897 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8fpt\" (UniqueName: \"kubernetes.io/projected/e6defdc7-3e4e-497d-933c-896f079bc208-kube-api-access-c8fpt\") pod \"collect-profiles-29320425-bm8wq\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.497135 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:00 crc kubenswrapper[4810]: I0930 09:45:00.993116 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq"] Sep 30 09:45:01 crc kubenswrapper[4810]: W0930 09:45:01.000112 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6defdc7_3e4e_497d_933c_896f079bc208.slice/crio-50d6e757f07d090c467e27980bc744a2ba400a0a112d8fbcc8336c05a601f47d WatchSource:0}: Error finding container 50d6e757f07d090c467e27980bc744a2ba400a0a112d8fbcc8336c05a601f47d: Status 404 returned error can't find the container with id 50d6e757f07d090c467e27980bc744a2ba400a0a112d8fbcc8336c05a601f47d Sep 30 09:45:01 crc kubenswrapper[4810]: I0930 09:45:01.054018 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" event={"ID":"e6defdc7-3e4e-497d-933c-896f079bc208","Type":"ContainerStarted","Data":"50d6e757f07d090c467e27980bc744a2ba400a0a112d8fbcc8336c05a601f47d"} Sep 30 09:45:02 crc kubenswrapper[4810]: I0930 09:45:02.064329 4810 generic.go:334] "Generic (PLEG): container finished" podID="e6defdc7-3e4e-497d-933c-896f079bc208" containerID="50f14fc2c7172d3707cd6d518f80e212d3b801b70ba5419c7cc3833afc80b2dc" exitCode=0 Sep 30 09:45:02 crc kubenswrapper[4810]: I0930 09:45:02.064664 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" event={"ID":"e6defdc7-3e4e-497d-933c-896f079bc208","Type":"ContainerDied","Data":"50f14fc2c7172d3707cd6d518f80e212d3b801b70ba5419c7cc3833afc80b2dc"} Sep 30 09:45:03 crc kubenswrapper[4810]: I0930 09:45:03.420921 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:03 crc kubenswrapper[4810]: I0930 09:45:03.489555 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6defdc7-3e4e-497d-933c-896f079bc208-config-volume\") pod \"e6defdc7-3e4e-497d-933c-896f079bc208\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " Sep 30 09:45:03 crc kubenswrapper[4810]: I0930 09:45:03.489789 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6defdc7-3e4e-497d-933c-896f079bc208-secret-volume\") pod \"e6defdc7-3e4e-497d-933c-896f079bc208\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " Sep 30 09:45:03 crc kubenswrapper[4810]: I0930 09:45:03.489955 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8fpt\" (UniqueName: \"kubernetes.io/projected/e6defdc7-3e4e-497d-933c-896f079bc208-kube-api-access-c8fpt\") pod \"e6defdc7-3e4e-497d-933c-896f079bc208\" (UID: \"e6defdc7-3e4e-497d-933c-896f079bc208\") " Sep 30 09:45:03 crc kubenswrapper[4810]: I0930 09:45:03.490340 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6defdc7-3e4e-497d-933c-896f079bc208-config-volume" (OuterVolumeSpecName: "config-volume") pod "e6defdc7-3e4e-497d-933c-896f079bc208" (UID: "e6defdc7-3e4e-497d-933c-896f079bc208"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 09:45:03 crc kubenswrapper[4810]: I0930 09:45:03.490511 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6defdc7-3e4e-497d-933c-896f079bc208-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 09:45:03 crc kubenswrapper[4810]: I0930 09:45:03.499567 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6defdc7-3e4e-497d-933c-896f079bc208-kube-api-access-c8fpt" (OuterVolumeSpecName: "kube-api-access-c8fpt") pod "e6defdc7-3e4e-497d-933c-896f079bc208" (UID: "e6defdc7-3e4e-497d-933c-896f079bc208"). InnerVolumeSpecName "kube-api-access-c8fpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:45:03 crc kubenswrapper[4810]: I0930 09:45:03.502986 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6defdc7-3e4e-497d-933c-896f079bc208-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e6defdc7-3e4e-497d-933c-896f079bc208" (UID: "e6defdc7-3e4e-497d-933c-896f079bc208"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 09:45:03 crc kubenswrapper[4810]: I0930 09:45:03.592597 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6defdc7-3e4e-497d-933c-896f079bc208-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 09:45:03 crc kubenswrapper[4810]: I0930 09:45:03.592633 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8fpt\" (UniqueName: \"kubernetes.io/projected/e6defdc7-3e4e-497d-933c-896f079bc208-kube-api-access-c8fpt\") on node \"crc\" DevicePath \"\"" Sep 30 09:45:04 crc kubenswrapper[4810]: I0930 09:45:04.097177 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" event={"ID":"e6defdc7-3e4e-497d-933c-896f079bc208","Type":"ContainerDied","Data":"50d6e757f07d090c467e27980bc744a2ba400a0a112d8fbcc8336c05a601f47d"} Sep 30 09:45:04 crc kubenswrapper[4810]: I0930 09:45:04.097525 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50d6e757f07d090c467e27980bc744a2ba400a0a112d8fbcc8336c05a601f47d" Sep 30 09:45:04 crc kubenswrapper[4810]: I0930 09:45:04.097226 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320425-bm8wq" Sep 30 09:45:04 crc kubenswrapper[4810]: I0930 09:45:04.494094 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h"] Sep 30 09:45:04 crc kubenswrapper[4810]: I0930 09:45:04.502692 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320380-fkg4h"] Sep 30 09:45:05 crc kubenswrapper[4810]: I0930 09:45:05.317808 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d7cf5d7-ce91-4c1b-b025-2edb9e25353c" path="/var/lib/kubelet/pods/6d7cf5d7-ce91-4c1b-b025-2edb9e25353c/volumes" Sep 30 09:45:09 crc kubenswrapper[4810]: I0930 09:45:09.154208 4810 generic.go:334] "Generic (PLEG): container finished" podID="281ea86c-1429-40e5-9e26-2f06551a023c" containerID="57622db53c54a7692bfe9230fe7b91e7386251715175014a2bff2e21c3fa02bc" exitCode=0 Sep 30 09:45:09 crc kubenswrapper[4810]: I0930 09:45:09.154311 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/crc-debug-mdwmv" event={"ID":"281ea86c-1429-40e5-9e26-2f06551a023c","Type":"ContainerDied","Data":"57622db53c54a7692bfe9230fe7b91e7386251715175014a2bff2e21c3fa02bc"} Sep 30 09:45:10 crc kubenswrapper[4810]: I0930 09:45:10.263553 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-mdwmv" Sep 30 09:45:10 crc kubenswrapper[4810]: I0930 09:45:10.299505 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pdntm/crc-debug-mdwmv"] Sep 30 09:45:10 crc kubenswrapper[4810]: I0930 09:45:10.320951 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pdntm/crc-debug-mdwmv"] Sep 30 09:45:10 crc kubenswrapper[4810]: I0930 09:45:10.321394 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:45:10 crc kubenswrapper[4810]: E0930 09:45:10.321978 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:45:10 crc kubenswrapper[4810]: I0930 09:45:10.333220 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8p8b\" (UniqueName: \"kubernetes.io/projected/281ea86c-1429-40e5-9e26-2f06551a023c-kube-api-access-b8p8b\") pod \"281ea86c-1429-40e5-9e26-2f06551a023c\" (UID: \"281ea86c-1429-40e5-9e26-2f06551a023c\") " Sep 30 09:45:10 crc kubenswrapper[4810]: I0930 09:45:10.333491 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/281ea86c-1429-40e5-9e26-2f06551a023c-host\") pod \"281ea86c-1429-40e5-9e26-2f06551a023c\" (UID: \"281ea86c-1429-40e5-9e26-2f06551a023c\") " Sep 30 09:45:10 crc kubenswrapper[4810]: I0930 09:45:10.333890 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/281ea86c-1429-40e5-9e26-2f06551a023c-host" (OuterVolumeSpecName: "host") pod "281ea86c-1429-40e5-9e26-2f06551a023c" (UID: "281ea86c-1429-40e5-9e26-2f06551a023c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 09:45:10 crc kubenswrapper[4810]: I0930 09:45:10.341719 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/281ea86c-1429-40e5-9e26-2f06551a023c-kube-api-access-b8p8b" (OuterVolumeSpecName: "kube-api-access-b8p8b") pod "281ea86c-1429-40e5-9e26-2f06551a023c" (UID: "281ea86c-1429-40e5-9e26-2f06551a023c"). InnerVolumeSpecName "kube-api-access-b8p8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:45:10 crc kubenswrapper[4810]: I0930 09:45:10.436722 4810 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/281ea86c-1429-40e5-9e26-2f06551a023c-host\") on node \"crc\" DevicePath \"\"" Sep 30 09:45:10 crc kubenswrapper[4810]: I0930 09:45:10.436763 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8p8b\" (UniqueName: \"kubernetes.io/projected/281ea86c-1429-40e5-9e26-2f06551a023c-kube-api-access-b8p8b\") on node \"crc\" DevicePath \"\"" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.178668 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50cd66f8ed8f692c534af9aa555be051d87b9137fa185098d814c583804a9dea" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.178776 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-mdwmv" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.329806 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="281ea86c-1429-40e5-9e26-2f06551a023c" path="/var/lib/kubelet/pods/281ea86c-1429-40e5-9e26-2f06551a023c/volumes" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.515573 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pdntm/crc-debug-f8vmw"] Sep 30 09:45:11 crc kubenswrapper[4810]: E0930 09:45:11.515937 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="281ea86c-1429-40e5-9e26-2f06551a023c" containerName="container-00" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.515953 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="281ea86c-1429-40e5-9e26-2f06551a023c" containerName="container-00" Sep 30 09:45:11 crc kubenswrapper[4810]: E0930 09:45:11.515976 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6defdc7-3e4e-497d-933c-896f079bc208" containerName="collect-profiles" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.515984 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6defdc7-3e4e-497d-933c-896f079bc208" containerName="collect-profiles" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.516186 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="281ea86c-1429-40e5-9e26-2f06551a023c" containerName="container-00" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.516207 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6defdc7-3e4e-497d-933c-896f079bc208" containerName="collect-profiles" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.516915 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-f8vmw" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.520292 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pdntm"/"default-dockercfg-pllwj" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.661206 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdcpw\" (UniqueName: \"kubernetes.io/projected/bea8809b-eaa9-46a3-84f7-081532e7fb8f-kube-api-access-zdcpw\") pod \"crc-debug-f8vmw\" (UID: \"bea8809b-eaa9-46a3-84f7-081532e7fb8f\") " pod="openshift-must-gather-pdntm/crc-debug-f8vmw" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.661371 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bea8809b-eaa9-46a3-84f7-081532e7fb8f-host\") pod \"crc-debug-f8vmw\" (UID: \"bea8809b-eaa9-46a3-84f7-081532e7fb8f\") " pod="openshift-must-gather-pdntm/crc-debug-f8vmw" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.764027 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdcpw\" (UniqueName: \"kubernetes.io/projected/bea8809b-eaa9-46a3-84f7-081532e7fb8f-kube-api-access-zdcpw\") pod \"crc-debug-f8vmw\" (UID: \"bea8809b-eaa9-46a3-84f7-081532e7fb8f\") " pod="openshift-must-gather-pdntm/crc-debug-f8vmw" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.764100 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bea8809b-eaa9-46a3-84f7-081532e7fb8f-host\") pod \"crc-debug-f8vmw\" (UID: \"bea8809b-eaa9-46a3-84f7-081532e7fb8f\") " pod="openshift-must-gather-pdntm/crc-debug-f8vmw" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.764304 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bea8809b-eaa9-46a3-84f7-081532e7fb8f-host\") pod \"crc-debug-f8vmw\" (UID: \"bea8809b-eaa9-46a3-84f7-081532e7fb8f\") " pod="openshift-must-gather-pdntm/crc-debug-f8vmw" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.783156 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdcpw\" (UniqueName: \"kubernetes.io/projected/bea8809b-eaa9-46a3-84f7-081532e7fb8f-kube-api-access-zdcpw\") pod \"crc-debug-f8vmw\" (UID: \"bea8809b-eaa9-46a3-84f7-081532e7fb8f\") " pod="openshift-must-gather-pdntm/crc-debug-f8vmw" Sep 30 09:45:11 crc kubenswrapper[4810]: I0930 09:45:11.844460 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-f8vmw" Sep 30 09:45:12 crc kubenswrapper[4810]: I0930 09:45:12.189997 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/crc-debug-f8vmw" event={"ID":"bea8809b-eaa9-46a3-84f7-081532e7fb8f","Type":"ContainerStarted","Data":"dedcbbb66df6bf7b2242960a1e62a57499b8ed93542c8b1041e885fa4651c6ee"} Sep 30 09:45:12 crc kubenswrapper[4810]: I0930 09:45:12.190053 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/crc-debug-f8vmw" event={"ID":"bea8809b-eaa9-46a3-84f7-081532e7fb8f","Type":"ContainerStarted","Data":"ceb2c2d86149f787c78b7fade91bb9cf655a1db3a6a4aed1f3854e0256a588d1"} Sep 30 09:45:12 crc kubenswrapper[4810]: I0930 09:45:12.213256 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pdntm/crc-debug-f8vmw" podStartSLOduration=1.213237461 podStartE2EDuration="1.213237461s" podCreationTimestamp="2025-09-30 09:45:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 09:45:12.201914232 +0000 UTC m=+6135.654113499" watchObservedRunningTime="2025-09-30 09:45:12.213237461 +0000 UTC m=+6135.665436728" Sep 30 09:45:13 crc kubenswrapper[4810]: I0930 09:45:13.200813 4810 generic.go:334] "Generic (PLEG): container finished" podID="bea8809b-eaa9-46a3-84f7-081532e7fb8f" containerID="dedcbbb66df6bf7b2242960a1e62a57499b8ed93542c8b1041e885fa4651c6ee" exitCode=0 Sep 30 09:45:13 crc kubenswrapper[4810]: I0930 09:45:13.200877 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/crc-debug-f8vmw" event={"ID":"bea8809b-eaa9-46a3-84f7-081532e7fb8f","Type":"ContainerDied","Data":"dedcbbb66df6bf7b2242960a1e62a57499b8ed93542c8b1041e885fa4651c6ee"} Sep 30 09:45:14 crc kubenswrapper[4810]: I0930 09:45:14.327916 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-f8vmw" Sep 30 09:45:14 crc kubenswrapper[4810]: I0930 09:45:14.419686 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bea8809b-eaa9-46a3-84f7-081532e7fb8f-host\") pod \"bea8809b-eaa9-46a3-84f7-081532e7fb8f\" (UID: \"bea8809b-eaa9-46a3-84f7-081532e7fb8f\") " Sep 30 09:45:14 crc kubenswrapper[4810]: I0930 09:45:14.419791 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdcpw\" (UniqueName: \"kubernetes.io/projected/bea8809b-eaa9-46a3-84f7-081532e7fb8f-kube-api-access-zdcpw\") pod \"bea8809b-eaa9-46a3-84f7-081532e7fb8f\" (UID: \"bea8809b-eaa9-46a3-84f7-081532e7fb8f\") " Sep 30 09:45:14 crc kubenswrapper[4810]: I0930 09:45:14.419915 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bea8809b-eaa9-46a3-84f7-081532e7fb8f-host" (OuterVolumeSpecName: "host") pod "bea8809b-eaa9-46a3-84f7-081532e7fb8f" (UID: "bea8809b-eaa9-46a3-84f7-081532e7fb8f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 09:45:14 crc kubenswrapper[4810]: I0930 09:45:14.420602 4810 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bea8809b-eaa9-46a3-84f7-081532e7fb8f-host\") on node \"crc\" DevicePath \"\"" Sep 30 09:45:14 crc kubenswrapper[4810]: I0930 09:45:14.429354 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bea8809b-eaa9-46a3-84f7-081532e7fb8f-kube-api-access-zdcpw" (OuterVolumeSpecName: "kube-api-access-zdcpw") pod "bea8809b-eaa9-46a3-84f7-081532e7fb8f" (UID: "bea8809b-eaa9-46a3-84f7-081532e7fb8f"). InnerVolumeSpecName "kube-api-access-zdcpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:45:14 crc kubenswrapper[4810]: I0930 09:45:14.524356 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdcpw\" (UniqueName: \"kubernetes.io/projected/bea8809b-eaa9-46a3-84f7-081532e7fb8f-kube-api-access-zdcpw\") on node \"crc\" DevicePath \"\"" Sep 30 09:45:15 crc kubenswrapper[4810]: I0930 09:45:15.219395 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/crc-debug-f8vmw" event={"ID":"bea8809b-eaa9-46a3-84f7-081532e7fb8f","Type":"ContainerDied","Data":"ceb2c2d86149f787c78b7fade91bb9cf655a1db3a6a4aed1f3854e0256a588d1"} Sep 30 09:45:15 crc kubenswrapper[4810]: I0930 09:45:15.219442 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-f8vmw" Sep 30 09:45:15 crc kubenswrapper[4810]: I0930 09:45:15.219448 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ceb2c2d86149f787c78b7fade91bb9cf655a1db3a6a4aed1f3854e0256a588d1" Sep 30 09:45:21 crc kubenswrapper[4810]: I0930 09:45:21.851090 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pdntm/crc-debug-f8vmw"] Sep 30 09:45:21 crc kubenswrapper[4810]: I0930 09:45:21.860605 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pdntm/crc-debug-f8vmw"] Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.041253 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pdntm/crc-debug-62gnr"] Sep 30 09:45:23 crc kubenswrapper[4810]: E0930 09:45:23.042071 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea8809b-eaa9-46a3-84f7-081532e7fb8f" containerName="container-00" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.042090 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea8809b-eaa9-46a3-84f7-081532e7fb8f" containerName="container-00" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.042356 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bea8809b-eaa9-46a3-84f7-081532e7fb8f" containerName="container-00" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.043214 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-62gnr" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.045660 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pdntm"/"default-dockercfg-pllwj" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.105780 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a37c74d-44d9-41c2-b7f6-1aff44aad906-host\") pod \"crc-debug-62gnr\" (UID: \"8a37c74d-44d9-41c2-b7f6-1aff44aad906\") " pod="openshift-must-gather-pdntm/crc-debug-62gnr" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.106058 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlrst\" (UniqueName: \"kubernetes.io/projected/8a37c74d-44d9-41c2-b7f6-1aff44aad906-kube-api-access-jlrst\") pod \"crc-debug-62gnr\" (UID: \"8a37c74d-44d9-41c2-b7f6-1aff44aad906\") " pod="openshift-must-gather-pdntm/crc-debug-62gnr" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.207882 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a37c74d-44d9-41c2-b7f6-1aff44aad906-host\") pod \"crc-debug-62gnr\" (UID: \"8a37c74d-44d9-41c2-b7f6-1aff44aad906\") " pod="openshift-must-gather-pdntm/crc-debug-62gnr" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.207993 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlrst\" (UniqueName: \"kubernetes.io/projected/8a37c74d-44d9-41c2-b7f6-1aff44aad906-kube-api-access-jlrst\") pod \"crc-debug-62gnr\" (UID: \"8a37c74d-44d9-41c2-b7f6-1aff44aad906\") " pod="openshift-must-gather-pdntm/crc-debug-62gnr" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.208067 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a37c74d-44d9-41c2-b7f6-1aff44aad906-host\") pod \"crc-debug-62gnr\" (UID: \"8a37c74d-44d9-41c2-b7f6-1aff44aad906\") " pod="openshift-must-gather-pdntm/crc-debug-62gnr" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.229845 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlrst\" (UniqueName: \"kubernetes.io/projected/8a37c74d-44d9-41c2-b7f6-1aff44aad906-kube-api-access-jlrst\") pod \"crc-debug-62gnr\" (UID: \"8a37c74d-44d9-41c2-b7f6-1aff44aad906\") " pod="openshift-must-gather-pdntm/crc-debug-62gnr" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.307187 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:45:23 crc kubenswrapper[4810]: E0930 09:45:23.307622 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.319687 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bea8809b-eaa9-46a3-84f7-081532e7fb8f" path="/var/lib/kubelet/pods/bea8809b-eaa9-46a3-84f7-081532e7fb8f/volumes" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.363451 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-62gnr" Sep 30 09:45:23 crc kubenswrapper[4810]: I0930 09:45:23.767279 4810 scope.go:117] "RemoveContainer" containerID="c28a79315907100a129348ab98b29ba1ed0e3503b4e473970cc164e2fcfd6b7d" Sep 30 09:45:24 crc kubenswrapper[4810]: I0930 09:45:24.306047 4810 generic.go:334] "Generic (PLEG): container finished" podID="8a37c74d-44d9-41c2-b7f6-1aff44aad906" containerID="c6f9defdb3139050c78b0c71f828e67876fed163a8ed168c03f80e9faf03416a" exitCode=0 Sep 30 09:45:24 crc kubenswrapper[4810]: I0930 09:45:24.306105 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/crc-debug-62gnr" event={"ID":"8a37c74d-44d9-41c2-b7f6-1aff44aad906","Type":"ContainerDied","Data":"c6f9defdb3139050c78b0c71f828e67876fed163a8ed168c03f80e9faf03416a"} Sep 30 09:45:24 crc kubenswrapper[4810]: I0930 09:45:24.306197 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/crc-debug-62gnr" event={"ID":"8a37c74d-44d9-41c2-b7f6-1aff44aad906","Type":"ContainerStarted","Data":"a47dfca0e931603efcd6d305e5efa77bcc471aa176461401d846a94961952952"} Sep 30 09:45:24 crc kubenswrapper[4810]: I0930 09:45:24.352326 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pdntm/crc-debug-62gnr"] Sep 30 09:45:24 crc kubenswrapper[4810]: I0930 09:45:24.364801 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pdntm/crc-debug-62gnr"] Sep 30 09:45:25 crc kubenswrapper[4810]: I0930 09:45:25.431556 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-62gnr" Sep 30 09:45:25 crc kubenswrapper[4810]: I0930 09:45:25.554918 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlrst\" (UniqueName: \"kubernetes.io/projected/8a37c74d-44d9-41c2-b7f6-1aff44aad906-kube-api-access-jlrst\") pod \"8a37c74d-44d9-41c2-b7f6-1aff44aad906\" (UID: \"8a37c74d-44d9-41c2-b7f6-1aff44aad906\") " Sep 30 09:45:25 crc kubenswrapper[4810]: I0930 09:45:25.555161 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a37c74d-44d9-41c2-b7f6-1aff44aad906-host\") pod \"8a37c74d-44d9-41c2-b7f6-1aff44aad906\" (UID: \"8a37c74d-44d9-41c2-b7f6-1aff44aad906\") " Sep 30 09:45:25 crc kubenswrapper[4810]: I0930 09:45:25.555719 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a37c74d-44d9-41c2-b7f6-1aff44aad906-host" (OuterVolumeSpecName: "host") pod "8a37c74d-44d9-41c2-b7f6-1aff44aad906" (UID: "8a37c74d-44d9-41c2-b7f6-1aff44aad906"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 09:45:25 crc kubenswrapper[4810]: I0930 09:45:25.562029 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a37c74d-44d9-41c2-b7f6-1aff44aad906-kube-api-access-jlrst" (OuterVolumeSpecName: "kube-api-access-jlrst") pod "8a37c74d-44d9-41c2-b7f6-1aff44aad906" (UID: "8a37c74d-44d9-41c2-b7f6-1aff44aad906"). InnerVolumeSpecName "kube-api-access-jlrst". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:45:25 crc kubenswrapper[4810]: I0930 09:45:25.657577 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlrst\" (UniqueName: \"kubernetes.io/projected/8a37c74d-44d9-41c2-b7f6-1aff44aad906-kube-api-access-jlrst\") on node \"crc\" DevicePath \"\"" Sep 30 09:45:25 crc kubenswrapper[4810]: I0930 09:45:25.657630 4810 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a37c74d-44d9-41c2-b7f6-1aff44aad906-host\") on node \"crc\" DevicePath \"\"" Sep 30 09:45:26 crc kubenswrapper[4810]: I0930 09:45:26.326821 4810 scope.go:117] "RemoveContainer" containerID="c6f9defdb3139050c78b0c71f828e67876fed163a8ed168c03f80e9faf03416a" Sep 30 09:45:26 crc kubenswrapper[4810]: I0930 09:45:26.326873 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/crc-debug-62gnr" Sep 30 09:45:26 crc kubenswrapper[4810]: I0930 09:45:26.507856 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/util/0.log" Sep 30 09:45:26 crc kubenswrapper[4810]: I0930 09:45:26.764517 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/pull/0.log" Sep 30 09:45:26 crc kubenswrapper[4810]: I0930 09:45:26.792883 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/pull/0.log" Sep 30 09:45:26 crc kubenswrapper[4810]: I0930 09:45:26.809603 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/util/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.000670 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/util/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.011816 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/pull/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.020009 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0de4ce1489fcacdd9c5879a788833a0fee1b9fa1f1b988fad5f7e9179dm88nv_5f1876f1-ed80-4dce-bdce-6a0c07a4f675/extract/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.267673 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-zzj8m_6ad79408-8dc2-4e7e-aee3-54e3778a344e/manager/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.272872 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-zzj8m_6ad79408-8dc2-4e7e-aee3-54e3778a344e/kube-rbac-proxy/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.317695 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a37c74d-44d9-41c2-b7f6-1aff44aad906" path="/var/lib/kubelet/pods/8a37c74d-44d9-41c2-b7f6-1aff44aad906/volumes" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.322780 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-6dhj8_982a5d6f-5897-40d0-88a4-40a9f1ae9a5f/kube-rbac-proxy/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.516050 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-zfcrn_02d77694-4bb0-4735-94c1-a3314ee634b9/kube-rbac-proxy/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.554245 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-6dhj8_982a5d6f-5897-40d0-88a4-40a9f1ae9a5f/manager/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.567737 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-zfcrn_02d77694-4bb0-4735-94c1-a3314ee634b9/manager/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.739984 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-9wfgs_da3d67ab-80be-486a-bc4d-d36ae6eb4756/kube-rbac-proxy/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.833047 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-9wfgs_da3d67ab-80be-486a-bc4d-d36ae6eb4756/manager/0.log" Sep 30 09:45:27 crc kubenswrapper[4810]: I0930 09:45:27.947836 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-npvdf_e801f35d-5d9f-48df-9a66-dd11ca14d26c/kube-rbac-proxy/0.log" Sep 30 09:45:28 crc kubenswrapper[4810]: I0930 09:45:28.077941 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-npvdf_e801f35d-5d9f-48df-9a66-dd11ca14d26c/manager/0.log" Sep 30 09:45:28 crc kubenswrapper[4810]: I0930 09:45:28.163897 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-g76mw_09ab9971-634a-40dd-88df-c1093fb66766/kube-rbac-proxy/0.log" Sep 30 09:45:28 crc kubenswrapper[4810]: I0930 09:45:28.233010 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-g76mw_09ab9971-634a-40dd-88df-c1093fb66766/manager/0.log" Sep 30 09:45:28 crc kubenswrapper[4810]: I0930 09:45:28.438181 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-v9pjf_8b3a8007-4969-4c86-a03a-ce633b3c31aa/kube-rbac-proxy/0.log" Sep 30 09:45:28 crc kubenswrapper[4810]: I0930 09:45:28.530990 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-v9pjf_8b3a8007-4969-4c86-a03a-ce633b3c31aa/manager/0.log" Sep 30 09:45:28 crc kubenswrapper[4810]: I0930 09:45:28.579354 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-hd4x4_8c5f7678-0f63-4323-910c-227040fdac66/kube-rbac-proxy/0.log" Sep 30 09:45:28 crc kubenswrapper[4810]: I0930 09:45:28.688016 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-hd4x4_8c5f7678-0f63-4323-910c-227040fdac66/manager/0.log" Sep 30 09:45:28 crc kubenswrapper[4810]: I0930 09:45:28.794541 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-vz9pd_52a2d49f-0736-4cac-8363-1a71fa6c1dd4/kube-rbac-proxy/0.log" Sep 30 09:45:28 crc kubenswrapper[4810]: I0930 09:45:28.834413 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-vz9pd_52a2d49f-0736-4cac-8363-1a71fa6c1dd4/manager/0.log" Sep 30 09:45:29 crc kubenswrapper[4810]: I0930 09:45:29.010834 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-8bwwd_df615184-edc5-4d93-b0ae-b414d9eb0ce1/kube-rbac-proxy/0.log" Sep 30 09:45:29 crc kubenswrapper[4810]: I0930 09:45:29.059539 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-8bwwd_df615184-edc5-4d93-b0ae-b414d9eb0ce1/manager/0.log" Sep 30 09:45:29 crc kubenswrapper[4810]: I0930 09:45:29.168107 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-wnngc_8a8a969d-2c5d-4229-b01d-cabcf3aec35c/kube-rbac-proxy/0.log" Sep 30 09:45:29 crc kubenswrapper[4810]: I0930 09:45:29.229281 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-wnngc_8a8a969d-2c5d-4229-b01d-cabcf3aec35c/manager/0.log" Sep 30 09:45:29 crc kubenswrapper[4810]: I0930 09:45:29.520763 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-t6wwm_ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e/kube-rbac-proxy/0.log" Sep 30 09:45:29 crc kubenswrapper[4810]: I0930 09:45:29.630893 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-t6wwm_ed64e30e-e940-4d3b-8b26-ebe49b0ddf3e/manager/0.log" Sep 30 09:45:29 crc kubenswrapper[4810]: I0930 09:45:29.709661 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-v58wt_b32751fb-95c0-4295-937a-ce7f82562863/kube-rbac-proxy/0.log" Sep 30 09:45:29 crc kubenswrapper[4810]: I0930 09:45:29.881332 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-v58wt_b32751fb-95c0-4295-937a-ce7f82562863/manager/0.log" Sep 30 09:45:29 crc kubenswrapper[4810]: I0930 09:45:29.948858 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-sndbm_0f34a7d1-aa5e-4b55-a75d-5366e196e742/kube-rbac-proxy/0.log" Sep 30 09:45:29 crc kubenswrapper[4810]: I0930 09:45:29.997672 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-sndbm_0f34a7d1-aa5e-4b55-a75d-5366e196e742/manager/0.log" Sep 30 09:45:30 crc kubenswrapper[4810]: I0930 09:45:30.142515 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-9gsjq_4e9c80fa-52fc-41db-a0be-babb0bd14b56/kube-rbac-proxy/0.log" Sep 30 09:45:30 crc kubenswrapper[4810]: I0930 09:45:30.212073 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-9gsjq_4e9c80fa-52fc-41db-a0be-babb0bd14b56/manager/0.log" Sep 30 09:45:30 crc kubenswrapper[4810]: I0930 09:45:30.241746 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-64647d55fc-6z69x_3df3a4ac-fc49-404f-aaba-cc8c053365c4/kube-rbac-proxy/0.log" Sep 30 09:45:30 crc kubenswrapper[4810]: I0930 09:45:30.469260 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6d66d9d974-d8fdr_154d7384-4a63-4a16-8400-903c4fe9eb8b/kube-rbac-proxy/0.log" Sep 30 09:45:30 crc kubenswrapper[4810]: I0930 09:45:30.632584 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-fqg48_75091a5f-7967-4317-90b6-e195f197ecdb/registry-server/0.log" Sep 30 09:45:30 crc kubenswrapper[4810]: I0930 09:45:30.692594 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6d66d9d974-d8fdr_154d7384-4a63-4a16-8400-903c4fe9eb8b/operator/0.log" Sep 30 09:45:30 crc kubenswrapper[4810]: I0930 09:45:30.724144 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-s785j_24044399-d033-436a-ad32-1e1aefe76626/kube-rbac-proxy/0.log" Sep 30 09:45:30 crc kubenswrapper[4810]: I0930 09:45:30.979300 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-t24hl_11f06c30-1cef-40f3-a493-8be235930b71/kube-rbac-proxy/0.log" Sep 30 09:45:30 crc kubenswrapper[4810]: I0930 09:45:30.991114 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-s785j_24044399-d033-436a-ad32-1e1aefe76626/manager/0.log" Sep 30 09:45:31 crc kubenswrapper[4810]: I0930 09:45:31.091864 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-t24hl_11f06c30-1cef-40f3-a493-8be235930b71/manager/0.log" Sep 30 09:45:31 crc kubenswrapper[4810]: I0930 09:45:31.308235 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-79d8469568-js8pw_bf4551e1-3981-407f-ba07-bff4db2b6d6d/operator/0.log" Sep 30 09:45:31 crc kubenswrapper[4810]: I0930 09:45:31.376390 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-tc2wc_3bbbc1cd-370d-468f-8985-cdf9219ca326/kube-rbac-proxy/0.log" Sep 30 09:45:31 crc kubenswrapper[4810]: I0930 09:45:31.470997 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-tc2wc_3bbbc1cd-370d-468f-8985-cdf9219ca326/manager/0.log" Sep 30 09:45:31 crc kubenswrapper[4810]: I0930 09:45:31.689982 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-g8ls7_0cc20b2d-4504-4f46-b0a0-5aee4b2201a3/kube-rbac-proxy/0.log" Sep 30 09:45:31 crc kubenswrapper[4810]: I0930 09:45:31.716817 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-64647d55fc-6z69x_3df3a4ac-fc49-404f-aaba-cc8c053365c4/manager/0.log" Sep 30 09:45:31 crc kubenswrapper[4810]: I0930 09:45:31.732354 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-g8s55_7fe8706c-6dfa-4a2f-876b-4cded24209bf/kube-rbac-proxy/0.log" Sep 30 09:45:31 crc kubenswrapper[4810]: I0930 09:45:31.924297 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-g8s55_7fe8706c-6dfa-4a2f-876b-4cded24209bf/manager/0.log" Sep 30 09:45:31 crc kubenswrapper[4810]: I0930 09:45:31.941836 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-g8ls7_0cc20b2d-4504-4f46-b0a0-5aee4b2201a3/manager/0.log" Sep 30 09:45:31 crc kubenswrapper[4810]: I0930 09:45:31.953858 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-59446d9d9-kb958_b63222b9-1cc3-4a55-a498-ce538cea80b9/kube-rbac-proxy/0.log" Sep 30 09:45:32 crc kubenswrapper[4810]: I0930 09:45:32.117487 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-59446d9d9-kb958_b63222b9-1cc3-4a55-a498-ce538cea80b9/manager/0.log" Sep 30 09:45:35 crc kubenswrapper[4810]: I0930 09:45:35.307463 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:45:35 crc kubenswrapper[4810]: E0930 09:45:35.308331 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:45:47 crc kubenswrapper[4810]: I0930 09:45:47.315771 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:45:47 crc kubenswrapper[4810]: E0930 09:45:47.316617 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:45:48 crc kubenswrapper[4810]: I0930 09:45:48.314339 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-p7cc4_00e1ebdd-3664-401d-b7bf-a9f786d9b6d2/control-plane-machine-set-operator/0.log" Sep 30 09:45:48 crc kubenswrapper[4810]: I0930 09:45:48.514648 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nj7lz_f7201c6b-de6b-4c4e-b20b-90c59d0db7ac/kube-rbac-proxy/0.log" Sep 30 09:45:48 crc kubenswrapper[4810]: I0930 09:45:48.556046 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nj7lz_f7201c6b-de6b-4c4e-b20b-90c59d0db7ac/machine-api-operator/0.log" Sep 30 09:45:58 crc kubenswrapper[4810]: I0930 09:45:58.307287 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:45:58 crc kubenswrapper[4810]: E0930 09:45:58.308166 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:46:00 crc kubenswrapper[4810]: I0930 09:46:00.435776 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-bdbrw_14ff18f7-0dac-4905-a94e-5a687ad05c8a/cert-manager-controller/0.log" Sep 30 09:46:00 crc kubenswrapper[4810]: I0930 09:46:00.613735 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-vbt4m_28b9d5cb-a422-4629-ba55-8bcd92f69971/cert-manager-cainjector/0.log" Sep 30 09:46:00 crc kubenswrapper[4810]: I0930 09:46:00.672870 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-wdstp_8090d644-f567-428e-a3b4-0d529d9bccb9/cert-manager-webhook/0.log" Sep 30 09:46:09 crc kubenswrapper[4810]: I0930 09:46:09.307560 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:46:09 crc kubenswrapper[4810]: E0930 09:46:09.308566 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:46:12 crc kubenswrapper[4810]: I0930 09:46:12.590172 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-pdhp7_b672b0bf-c5cc-467a-9014-99a3269488e8/nmstate-console-plugin/0.log" Sep 30 09:46:12 crc kubenswrapper[4810]: I0930 09:46:12.809567 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-w4dnr_b8df67fa-d671-4174-8306-61127ae22268/nmstate-handler/0.log" Sep 30 09:46:12 crc kubenswrapper[4810]: I0930 09:46:12.860636 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-qzvw4_303e8028-1bcf-4fa7-8314-2ed52bec230a/kube-rbac-proxy/0.log" Sep 30 09:46:12 crc kubenswrapper[4810]: I0930 09:46:12.911446 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-qzvw4_303e8028-1bcf-4fa7-8314-2ed52bec230a/nmstate-metrics/0.log" Sep 30 09:46:13 crc kubenswrapper[4810]: I0930 09:46:13.097531 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-bk866_5a1ec2d4-00e6-4db8-95c0-3079d24f18a8/nmstate-operator/0.log" Sep 30 09:46:13 crc kubenswrapper[4810]: I0930 09:46:13.136213 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-h2zdm_8dbcd62d-c15f-4d23-ac4a-80c8eda6dfcb/nmstate-webhook/0.log" Sep 30 09:46:22 crc kubenswrapper[4810]: I0930 09:46:22.306666 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:46:22 crc kubenswrapper[4810]: E0930 09:46:22.307444 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:46:27 crc kubenswrapper[4810]: I0930 09:46:27.339772 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-qr7xw_4f9b2513-5122-4757-b48c-4f0e8c9aaf05/kube-rbac-proxy/0.log" Sep 30 09:46:27 crc kubenswrapper[4810]: I0930 09:46:27.506915 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-qr7xw_4f9b2513-5122-4757-b48c-4f0e8c9aaf05/controller/0.log" Sep 30 09:46:27 crc kubenswrapper[4810]: I0930 09:46:27.591444 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-2cfm4_ad0bc482-465c-433b-bd10-82ed2f79b6ad/frr-k8s-webhook-server/0.log" Sep 30 09:46:27 crc kubenswrapper[4810]: I0930 09:46:27.703029 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-frr-files/0.log" Sep 30 09:46:27 crc kubenswrapper[4810]: I0930 09:46:27.868240 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-reloader/0.log" Sep 30 09:46:27 crc kubenswrapper[4810]: I0930 09:46:27.904615 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-metrics/0.log" Sep 30 09:46:27 crc kubenswrapper[4810]: I0930 09:46:27.916389 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-reloader/0.log" Sep 30 09:46:27 crc kubenswrapper[4810]: I0930 09:46:27.937088 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-frr-files/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.085999 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-frr-files/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.098065 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-reloader/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.175722 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-metrics/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.186155 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-metrics/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.315367 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-frr-files/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.325519 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-reloader/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.392634 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/cp-metrics/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.397986 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/controller/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.556665 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/frr-metrics/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.594763 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/kube-rbac-proxy/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.677477 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/kube-rbac-proxy-frr/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.782774 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/reloader/0.log" Sep 30 09:46:28 crc kubenswrapper[4810]: I0930 09:46:28.946923 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-74b66975c5-rtrd5_9710139d-e7f3-4921-b84a-4b6525891f99/manager/0.log" Sep 30 09:46:29 crc kubenswrapper[4810]: I0930 09:46:29.048189 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-768c8b4c95-d2v27_9d49935b-5628-43f0-b527-18d3b55aaa3a/webhook-server/0.log" Sep 30 09:46:29 crc kubenswrapper[4810]: I0930 09:46:29.242900 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-gjfhs_2f29bb1f-4702-4fa8-8a9e-692047b356f9/kube-rbac-proxy/0.log" Sep 30 09:46:29 crc kubenswrapper[4810]: I0930 09:46:29.875460 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-gjfhs_2f29bb1f-4702-4fa8-8a9e-692047b356f9/speaker/0.log" Sep 30 09:46:30 crc kubenswrapper[4810]: I0930 09:46:30.391500 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wfqtl_4debe3c7-7058-4f37-863c-0c0e57818c86/frr/0.log" Sep 30 09:46:34 crc kubenswrapper[4810]: I0930 09:46:34.308460 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:46:34 crc kubenswrapper[4810]: E0930 09:46:34.310589 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.039111 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/util/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.229491 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/util/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.249142 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/pull/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.254138 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/pull/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.455075 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/util/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.458490 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/pull/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.478496 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchb2p9_df44374d-4490-461f-aeed-05fae8096705/extract/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.640209 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/util/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.800538 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/pull/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.812608 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/util/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.813499 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/pull/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.941718 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/util/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.961177 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/pull/0.log" Sep 30 09:46:41 crc kubenswrapper[4810]: I0930 09:46:41.973152 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dt2mqs_9a4af0f1-83d1-4384-a16b-cb281e582ecb/extract/0.log" Sep 30 09:46:42 crc kubenswrapper[4810]: I0930 09:46:42.087777 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-utilities/0.log" Sep 30 09:46:42 crc kubenswrapper[4810]: I0930 09:46:42.274576 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-utilities/0.log" Sep 30 09:46:42 crc kubenswrapper[4810]: I0930 09:46:42.279049 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-content/0.log" Sep 30 09:46:42 crc kubenswrapper[4810]: I0930 09:46:42.285687 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-content/0.log" Sep 30 09:46:42 crc kubenswrapper[4810]: I0930 09:46:42.462121 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-utilities/0.log" Sep 30 09:46:42 crc kubenswrapper[4810]: I0930 09:46:42.469584 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/extract-content/0.log" Sep 30 09:46:42 crc kubenswrapper[4810]: I0930 09:46:42.714008 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-utilities/0.log" Sep 30 09:46:42 crc kubenswrapper[4810]: I0930 09:46:42.950877 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-content/0.log" Sep 30 09:46:42 crc kubenswrapper[4810]: I0930 09:46:42.966905 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-content/0.log" Sep 30 09:46:42 crc kubenswrapper[4810]: I0930 09:46:42.999196 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-utilities/0.log" Sep 30 09:46:43 crc kubenswrapper[4810]: I0930 09:46:43.183199 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-content/0.log" Sep 30 09:46:43 crc kubenswrapper[4810]: I0930 09:46:43.191864 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/extract-utilities/0.log" Sep 30 09:46:43 crc kubenswrapper[4810]: I0930 09:46:43.416541 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/util/0.log" Sep 30 09:46:43 crc kubenswrapper[4810]: I0930 09:46:43.617024 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/util/0.log" Sep 30 09:46:43 crc kubenswrapper[4810]: I0930 09:46:43.682000 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/pull/0.log" Sep 30 09:46:43 crc kubenswrapper[4810]: I0930 09:46:43.762194 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-skf7g_0546fe85-ddc6-4afa-afce-8ce05abba16b/registry-server/0.log" Sep 30 09:46:43 crc kubenswrapper[4810]: I0930 09:46:43.848331 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/pull/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.044381 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/util/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.069729 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zl5x9_7adac989-68a1-4f65-b249-7f43d8bac8be/registry-server/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.107195 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/extract/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.107932 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96n8hw8_cee1f0d2-0013-4803-b758-f3efb50a80b5/pull/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.267084 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-txqd2_7aace4e5-3195-4ec0-b0cf-c44d85c1f5ba/marketplace-operator/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.271534 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-utilities/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.488345 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-utilities/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.488431 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-content/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.501033 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-content/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.659887 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-content/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.757734 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/extract-utilities/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.771049 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-utilities/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.862921 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zvbp7_4ae8468f-2d42-4b13-9dbe-f8558df495cc/registry-server/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.966172 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-content/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.977232 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-utilities/0.log" Sep 30 09:46:44 crc kubenswrapper[4810]: I0930 09:46:44.996923 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-content/0.log" Sep 30 09:46:45 crc kubenswrapper[4810]: I0930 09:46:45.202943 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-content/0.log" Sep 30 09:46:45 crc kubenswrapper[4810]: I0930 09:46:45.214873 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/extract-utilities/0.log" Sep 30 09:46:45 crc kubenswrapper[4810]: I0930 09:46:45.363947 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x7zhw_6b679059-88b3-4a0f-8623-cb5a2a2c0e82/registry-server/0.log" Sep 30 09:46:46 crc kubenswrapper[4810]: I0930 09:46:46.307389 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:46:46 crc kubenswrapper[4810]: E0930 09:46:46.307664 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:46:56 crc kubenswrapper[4810]: I0930 09:46:56.171925 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-9cddq_360a4c93-d93b-4caf-b11d-a2185a5179d3/prometheus-operator/0.log" Sep 30 09:46:56 crc kubenswrapper[4810]: I0930 09:46:56.392431 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c865dc79d-4th6p_e41015ca-c417-4076-9895-abff0488bfd7/prometheus-operator-admission-webhook/0.log" Sep 30 09:46:56 crc kubenswrapper[4810]: I0930 09:46:56.397978 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c865dc79d-7jqf8_2ea01519-d717-45ef-be26-b429eb2c6cdb/prometheus-operator-admission-webhook/0.log" Sep 30 09:46:56 crc kubenswrapper[4810]: I0930 09:46:56.580400 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-kj7xb_5f51c1dc-ac89-4a84-8897-d2e9770baec9/perses-operator/0.log" Sep 30 09:46:56 crc kubenswrapper[4810]: I0930 09:46:56.611934 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-nq2k6_0da3209d-75c8-4eda-97f4-b6cfeac48f62/operator/0.log" Sep 30 09:46:59 crc kubenswrapper[4810]: I0930 09:46:59.306324 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:46:59 crc kubenswrapper[4810]: E0930 09:46:59.306959 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:47:13 crc kubenswrapper[4810]: I0930 09:47:13.307214 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:47:13 crc kubenswrapper[4810]: E0930 09:47:13.308832 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:47:24 crc kubenswrapper[4810]: I0930 09:47:24.306668 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:47:24 crc kubenswrapper[4810]: E0930 09:47:24.307505 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:47:37 crc kubenswrapper[4810]: I0930 09:47:37.313247 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:47:37 crc kubenswrapper[4810]: E0930 09:47:37.313962 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:47:51 crc kubenswrapper[4810]: I0930 09:47:51.306912 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:47:51 crc kubenswrapper[4810]: E0930 09:47:51.308806 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.121152 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j98l2"] Sep 30 09:47:59 crc kubenswrapper[4810]: E0930 09:47:59.125744 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a37c74d-44d9-41c2-b7f6-1aff44aad906" containerName="container-00" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.125768 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a37c74d-44d9-41c2-b7f6-1aff44aad906" containerName="container-00" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.126040 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a37c74d-44d9-41c2-b7f6-1aff44aad906" containerName="container-00" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.127982 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.138820 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j98l2"] Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.259221 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-catalog-content\") pod \"redhat-operators-j98l2\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.259326 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txpsx\" (UniqueName: \"kubernetes.io/projected/233df517-89ff-4918-9298-4ab2ad819cc9-kube-api-access-txpsx\") pod \"redhat-operators-j98l2\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.259511 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-utilities\") pod \"redhat-operators-j98l2\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.361592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-catalog-content\") pod \"redhat-operators-j98l2\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.361671 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txpsx\" (UniqueName: \"kubernetes.io/projected/233df517-89ff-4918-9298-4ab2ad819cc9-kube-api-access-txpsx\") pod \"redhat-operators-j98l2\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.361767 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-utilities\") pod \"redhat-operators-j98l2\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.362118 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-catalog-content\") pod \"redhat-operators-j98l2\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.362197 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-utilities\") pod \"redhat-operators-j98l2\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.382064 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txpsx\" (UniqueName: \"kubernetes.io/projected/233df517-89ff-4918-9298-4ab2ad819cc9-kube-api-access-txpsx\") pod \"redhat-operators-j98l2\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.461190 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:47:59 crc kubenswrapper[4810]: I0930 09:47:59.943965 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j98l2"] Sep 30 09:48:00 crc kubenswrapper[4810]: I0930 09:48:00.847081 4810 generic.go:334] "Generic (PLEG): container finished" podID="233df517-89ff-4918-9298-4ab2ad819cc9" containerID="9937c37d20730eb15a8ec9da1a4a11f79e0df903bd86a0eca669139690503612" exitCode=0 Sep 30 09:48:00 crc kubenswrapper[4810]: I0930 09:48:00.847254 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j98l2" event={"ID":"233df517-89ff-4918-9298-4ab2ad819cc9","Type":"ContainerDied","Data":"9937c37d20730eb15a8ec9da1a4a11f79e0df903bd86a0eca669139690503612"} Sep 30 09:48:00 crc kubenswrapper[4810]: I0930 09:48:00.847568 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j98l2" event={"ID":"233df517-89ff-4918-9298-4ab2ad819cc9","Type":"ContainerStarted","Data":"2e71eae41487965199824e42d7db8a91b3f1c186347953b005cf442969f003d0"} Sep 30 09:48:00 crc kubenswrapper[4810]: I0930 09:48:00.850114 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 09:48:02 crc kubenswrapper[4810]: I0930 09:48:02.871072 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j98l2" event={"ID":"233df517-89ff-4918-9298-4ab2ad819cc9","Type":"ContainerStarted","Data":"4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16"} Sep 30 09:48:03 crc kubenswrapper[4810]: I0930 09:48:03.308304 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:48:03 crc kubenswrapper[4810]: E0930 09:48:03.308720 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:48:04 crc kubenswrapper[4810]: I0930 09:48:04.898754 4810 generic.go:334] "Generic (PLEG): container finished" podID="233df517-89ff-4918-9298-4ab2ad819cc9" containerID="4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16" exitCode=0 Sep 30 09:48:04 crc kubenswrapper[4810]: I0930 09:48:04.898815 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j98l2" event={"ID":"233df517-89ff-4918-9298-4ab2ad819cc9","Type":"ContainerDied","Data":"4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16"} Sep 30 09:48:05 crc kubenswrapper[4810]: I0930 09:48:05.911753 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j98l2" event={"ID":"233df517-89ff-4918-9298-4ab2ad819cc9","Type":"ContainerStarted","Data":"0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8"} Sep 30 09:48:05 crc kubenswrapper[4810]: I0930 09:48:05.941261 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j98l2" podStartSLOduration=2.215179298 podStartE2EDuration="6.941236088s" podCreationTimestamp="2025-09-30 09:47:59 +0000 UTC" firstStartedPulling="2025-09-30 09:48:00.849850952 +0000 UTC m=+6304.302050219" lastFinishedPulling="2025-09-30 09:48:05.575907742 +0000 UTC m=+6309.028107009" observedRunningTime="2025-09-30 09:48:05.93614024 +0000 UTC m=+6309.388339507" watchObservedRunningTime="2025-09-30 09:48:05.941236088 +0000 UTC m=+6309.393435355" Sep 30 09:48:09 crc kubenswrapper[4810]: I0930 09:48:09.462337 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:48:09 crc kubenswrapper[4810]: I0930 09:48:09.462611 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:48:10 crc kubenswrapper[4810]: I0930 09:48:10.516102 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j98l2" podUID="233df517-89ff-4918-9298-4ab2ad819cc9" containerName="registry-server" probeResult="failure" output=< Sep 30 09:48:10 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Sep 30 09:48:10 crc kubenswrapper[4810]: > Sep 30 09:48:15 crc kubenswrapper[4810]: I0930 09:48:15.306744 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:48:15 crc kubenswrapper[4810]: E0930 09:48:15.307839 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hk2k2_openshift-machine-config-operator(86194959-30fe-4da8-a64b-f5a849c2f73a)\"" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" Sep 30 09:48:19 crc kubenswrapper[4810]: I0930 09:48:19.506873 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:48:19 crc kubenswrapper[4810]: I0930 09:48:19.574083 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:48:19 crc kubenswrapper[4810]: I0930 09:48:19.756298 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j98l2"] Sep 30 09:48:21 crc kubenswrapper[4810]: I0930 09:48:21.076235 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j98l2" podUID="233df517-89ff-4918-9298-4ab2ad819cc9" containerName="registry-server" containerID="cri-o://0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8" gracePeriod=2 Sep 30 09:48:21 crc kubenswrapper[4810]: I0930 09:48:21.563394 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:48:21 crc kubenswrapper[4810]: I0930 09:48:21.747803 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txpsx\" (UniqueName: \"kubernetes.io/projected/233df517-89ff-4918-9298-4ab2ad819cc9-kube-api-access-txpsx\") pod \"233df517-89ff-4918-9298-4ab2ad819cc9\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " Sep 30 09:48:21 crc kubenswrapper[4810]: I0930 09:48:21.747932 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-catalog-content\") pod \"233df517-89ff-4918-9298-4ab2ad819cc9\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " Sep 30 09:48:21 crc kubenswrapper[4810]: I0930 09:48:21.748226 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-utilities\") pod \"233df517-89ff-4918-9298-4ab2ad819cc9\" (UID: \"233df517-89ff-4918-9298-4ab2ad819cc9\") " Sep 30 09:48:21 crc kubenswrapper[4810]: I0930 09:48:21.749674 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-utilities" (OuterVolumeSpecName: "utilities") pod "233df517-89ff-4918-9298-4ab2ad819cc9" (UID: "233df517-89ff-4918-9298-4ab2ad819cc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:48:21 crc kubenswrapper[4810]: I0930 09:48:21.753254 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/233df517-89ff-4918-9298-4ab2ad819cc9-kube-api-access-txpsx" (OuterVolumeSpecName: "kube-api-access-txpsx") pod "233df517-89ff-4918-9298-4ab2ad819cc9" (UID: "233df517-89ff-4918-9298-4ab2ad819cc9"). InnerVolumeSpecName "kube-api-access-txpsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:48:21 crc kubenswrapper[4810]: I0930 09:48:21.830143 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "233df517-89ff-4918-9298-4ab2ad819cc9" (UID: "233df517-89ff-4918-9298-4ab2ad819cc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:48:21 crc kubenswrapper[4810]: I0930 09:48:21.851167 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:48:21 crc kubenswrapper[4810]: I0930 09:48:21.851193 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/233df517-89ff-4918-9298-4ab2ad819cc9-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:48:21 crc kubenswrapper[4810]: I0930 09:48:21.851203 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txpsx\" (UniqueName: \"kubernetes.io/projected/233df517-89ff-4918-9298-4ab2ad819cc9-kube-api-access-txpsx\") on node \"crc\" DevicePath \"\"" Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.086618 4810 generic.go:334] "Generic (PLEG): container finished" podID="233df517-89ff-4918-9298-4ab2ad819cc9" containerID="0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8" exitCode=0 Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.086661 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j98l2" event={"ID":"233df517-89ff-4918-9298-4ab2ad819cc9","Type":"ContainerDied","Data":"0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8"} Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.086693 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j98l2" event={"ID":"233df517-89ff-4918-9298-4ab2ad819cc9","Type":"ContainerDied","Data":"2e71eae41487965199824e42d7db8a91b3f1c186347953b005cf442969f003d0"} Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.086709 4810 scope.go:117] "RemoveContainer" containerID="0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8" Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.086765 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j98l2" Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.125777 4810 scope.go:117] "RemoveContainer" containerID="4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16" Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.138985 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j98l2"] Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.155427 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j98l2"] Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.172960 4810 scope.go:117] "RemoveContainer" containerID="9937c37d20730eb15a8ec9da1a4a11f79e0df903bd86a0eca669139690503612" Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.218750 4810 scope.go:117] "RemoveContainer" containerID="0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8" Sep 30 09:48:22 crc kubenswrapper[4810]: E0930 09:48:22.219446 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8\": container with ID starting with 0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8 not found: ID does not exist" containerID="0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8" Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.219508 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8"} err="failed to get container status \"0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8\": rpc error: code = NotFound desc = could not find container \"0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8\": container with ID starting with 0b4f965609bea78994d5bbdb577a59e91c96f766c588e563433b9fac7598f6e8 not found: ID does not exist" Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.219540 4810 scope.go:117] "RemoveContainer" containerID="4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16" Sep 30 09:48:22 crc kubenswrapper[4810]: E0930 09:48:22.220069 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16\": container with ID starting with 4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16 not found: ID does not exist" containerID="4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16" Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.220108 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16"} err="failed to get container status \"4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16\": rpc error: code = NotFound desc = could not find container \"4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16\": container with ID starting with 4a61445b5261a478582f4ed6e4f39d97ad52bb853edcdcf49a9a0b07cccfcc16 not found: ID does not exist" Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.220137 4810 scope.go:117] "RemoveContainer" containerID="9937c37d20730eb15a8ec9da1a4a11f79e0df903bd86a0eca669139690503612" Sep 30 09:48:22 crc kubenswrapper[4810]: E0930 09:48:22.220678 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9937c37d20730eb15a8ec9da1a4a11f79e0df903bd86a0eca669139690503612\": container with ID starting with 9937c37d20730eb15a8ec9da1a4a11f79e0df903bd86a0eca669139690503612 not found: ID does not exist" containerID="9937c37d20730eb15a8ec9da1a4a11f79e0df903bd86a0eca669139690503612" Sep 30 09:48:22 crc kubenswrapper[4810]: I0930 09:48:22.220709 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9937c37d20730eb15a8ec9da1a4a11f79e0df903bd86a0eca669139690503612"} err="failed to get container status \"9937c37d20730eb15a8ec9da1a4a11f79e0df903bd86a0eca669139690503612\": rpc error: code = NotFound desc = could not find container \"9937c37d20730eb15a8ec9da1a4a11f79e0df903bd86a0eca669139690503612\": container with ID starting with 9937c37d20730eb15a8ec9da1a4a11f79e0df903bd86a0eca669139690503612 not found: ID does not exist" Sep 30 09:48:23 crc kubenswrapper[4810]: I0930 09:48:23.321592 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="233df517-89ff-4918-9298-4ab2ad819cc9" path="/var/lib/kubelet/pods/233df517-89ff-4918-9298-4ab2ad819cc9/volumes" Sep 30 09:48:30 crc kubenswrapper[4810]: I0930 09:48:30.306909 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:48:31 crc kubenswrapper[4810]: I0930 09:48:31.169016 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"19649ce9d9c97dcaa82fd170c48b84ab4529eb7ee9a0d6d010bcfc9f2397e8d8"} Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.397310 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-57v2x"] Sep 30 09:49:13 crc kubenswrapper[4810]: E0930 09:49:13.398902 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="233df517-89ff-4918-9298-4ab2ad819cc9" containerName="extract-utilities" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.398923 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="233df517-89ff-4918-9298-4ab2ad819cc9" containerName="extract-utilities" Sep 30 09:49:13 crc kubenswrapper[4810]: E0930 09:49:13.399033 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="233df517-89ff-4918-9298-4ab2ad819cc9" containerName="registry-server" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.399049 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="233df517-89ff-4918-9298-4ab2ad819cc9" containerName="registry-server" Sep 30 09:49:13 crc kubenswrapper[4810]: E0930 09:49:13.399081 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="233df517-89ff-4918-9298-4ab2ad819cc9" containerName="extract-content" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.399090 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="233df517-89ff-4918-9298-4ab2ad819cc9" containerName="extract-content" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.399650 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="233df517-89ff-4918-9298-4ab2ad819cc9" containerName="registry-server" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.405150 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.408912 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-57v2x"] Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.482829 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9fc86dc-5bcb-4d21-935b-bb114817406f-utilities\") pod \"redhat-marketplace-57v2x\" (UID: \"a9fc86dc-5bcb-4d21-935b-bb114817406f\") " pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.483175 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9fc86dc-5bcb-4d21-935b-bb114817406f-catalog-content\") pod \"redhat-marketplace-57v2x\" (UID: \"a9fc86dc-5bcb-4d21-935b-bb114817406f\") " pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.483324 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sd5z\" (UniqueName: \"kubernetes.io/projected/a9fc86dc-5bcb-4d21-935b-bb114817406f-kube-api-access-9sd5z\") pod \"redhat-marketplace-57v2x\" (UID: \"a9fc86dc-5bcb-4d21-935b-bb114817406f\") " pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.585081 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9fc86dc-5bcb-4d21-935b-bb114817406f-utilities\") pod \"redhat-marketplace-57v2x\" (UID: \"a9fc86dc-5bcb-4d21-935b-bb114817406f\") " pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.585378 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9fc86dc-5bcb-4d21-935b-bb114817406f-catalog-content\") pod \"redhat-marketplace-57v2x\" (UID: \"a9fc86dc-5bcb-4d21-935b-bb114817406f\") " pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.585545 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sd5z\" (UniqueName: \"kubernetes.io/projected/a9fc86dc-5bcb-4d21-935b-bb114817406f-kube-api-access-9sd5z\") pod \"redhat-marketplace-57v2x\" (UID: \"a9fc86dc-5bcb-4d21-935b-bb114817406f\") " pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.585602 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9fc86dc-5bcb-4d21-935b-bb114817406f-utilities\") pod \"redhat-marketplace-57v2x\" (UID: \"a9fc86dc-5bcb-4d21-935b-bb114817406f\") " pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.585698 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9fc86dc-5bcb-4d21-935b-bb114817406f-catalog-content\") pod \"redhat-marketplace-57v2x\" (UID: \"a9fc86dc-5bcb-4d21-935b-bb114817406f\") " pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.609552 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sd5z\" (UniqueName: \"kubernetes.io/projected/a9fc86dc-5bcb-4d21-935b-bb114817406f-kube-api-access-9sd5z\") pod \"redhat-marketplace-57v2x\" (UID: \"a9fc86dc-5bcb-4d21-935b-bb114817406f\") " pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:13 crc kubenswrapper[4810]: I0930 09:49:13.733997 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:14 crc kubenswrapper[4810]: I0930 09:49:14.282494 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-57v2x"] Sep 30 09:49:14 crc kubenswrapper[4810]: I0930 09:49:14.608979 4810 generic.go:334] "Generic (PLEG): container finished" podID="a9fc86dc-5bcb-4d21-935b-bb114817406f" containerID="d224ffd122d949a3b07a7767a1503dd21ab5462845341d8d2a9ef11483062e7c" exitCode=0 Sep 30 09:49:14 crc kubenswrapper[4810]: I0930 09:49:14.609096 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-57v2x" event={"ID":"a9fc86dc-5bcb-4d21-935b-bb114817406f","Type":"ContainerDied","Data":"d224ffd122d949a3b07a7767a1503dd21ab5462845341d8d2a9ef11483062e7c"} Sep 30 09:49:14 crc kubenswrapper[4810]: I0930 09:49:14.610209 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-57v2x" event={"ID":"a9fc86dc-5bcb-4d21-935b-bb114817406f","Type":"ContainerStarted","Data":"31f3a15feb0d6f2b0bd434780a985eb8ad37a033e2db2bdfabc2e625425487fd"} Sep 30 09:49:17 crc kubenswrapper[4810]: I0930 09:49:17.641141 4810 generic.go:334] "Generic (PLEG): container finished" podID="a9fc86dc-5bcb-4d21-935b-bb114817406f" containerID="185c69025a6d5236d46945ed3fafcf58a15e4b77019caf95cc354c50f39856bf" exitCode=0 Sep 30 09:49:17 crc kubenswrapper[4810]: I0930 09:49:17.641185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-57v2x" event={"ID":"a9fc86dc-5bcb-4d21-935b-bb114817406f","Type":"ContainerDied","Data":"185c69025a6d5236d46945ed3fafcf58a15e4b77019caf95cc354c50f39856bf"} Sep 30 09:49:18 crc kubenswrapper[4810]: I0930 09:49:18.653438 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-57v2x" event={"ID":"a9fc86dc-5bcb-4d21-935b-bb114817406f","Type":"ContainerStarted","Data":"791e0eba565d7251dbc9260b46c09d8fe0a633b5f97da1cb511df10e777fac33"} Sep 30 09:49:20 crc kubenswrapper[4810]: I0930 09:49:20.672756 4810 generic.go:334] "Generic (PLEG): container finished" podID="75f3fd61-26a0-4cb9-bdba-144e953d81e8" containerID="fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0" exitCode=0 Sep 30 09:49:20 crc kubenswrapper[4810]: I0930 09:49:20.672834 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pdntm/must-gather-2vvzs" event={"ID":"75f3fd61-26a0-4cb9-bdba-144e953d81e8","Type":"ContainerDied","Data":"fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0"} Sep 30 09:49:20 crc kubenswrapper[4810]: I0930 09:49:20.674154 4810 scope.go:117] "RemoveContainer" containerID="fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0" Sep 30 09:49:20 crc kubenswrapper[4810]: I0930 09:49:20.687383 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-57v2x" podStartSLOduration=4.193635398 podStartE2EDuration="7.687362867s" podCreationTimestamp="2025-09-30 09:49:13 +0000 UTC" firstStartedPulling="2025-09-30 09:49:14.612561868 +0000 UTC m=+6378.064761135" lastFinishedPulling="2025-09-30 09:49:18.106289327 +0000 UTC m=+6381.558488604" observedRunningTime="2025-09-30 09:49:18.682917262 +0000 UTC m=+6382.135116529" watchObservedRunningTime="2025-09-30 09:49:20.687362867 +0000 UTC m=+6384.139562134" Sep 30 09:49:20 crc kubenswrapper[4810]: I0930 09:49:20.761219 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pdntm_must-gather-2vvzs_75f3fd61-26a0-4cb9-bdba-144e953d81e8/gather/0.log" Sep 30 09:49:23 crc kubenswrapper[4810]: I0930 09:49:23.734302 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:23 crc kubenswrapper[4810]: I0930 09:49:23.734872 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:23 crc kubenswrapper[4810]: I0930 09:49:23.782892 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:23 crc kubenswrapper[4810]: I0930 09:49:23.922673 4810 scope.go:117] "RemoveContainer" containerID="57622db53c54a7692bfe9230fe7b91e7386251715175014a2bff2e21c3fa02bc" Sep 30 09:49:24 crc kubenswrapper[4810]: I0930 09:49:24.761409 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-57v2x" Sep 30 09:49:24 crc kubenswrapper[4810]: I0930 09:49:24.837966 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-57v2x"] Sep 30 09:49:24 crc kubenswrapper[4810]: I0930 09:49:24.869027 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvbp7"] Sep 30 09:49:24 crc kubenswrapper[4810]: I0930 09:49:24.869353 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zvbp7" podUID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" containerName="registry-server" containerID="cri-o://c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6" gracePeriod=2 Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.354126 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.523217 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-catalog-content\") pod \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.523325 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-utilities\") pod \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.523455 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdtt2\" (UniqueName: \"kubernetes.io/projected/4ae8468f-2d42-4b13-9dbe-f8558df495cc-kube-api-access-jdtt2\") pod \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\" (UID: \"4ae8468f-2d42-4b13-9dbe-f8558df495cc\") " Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.524293 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-utilities" (OuterVolumeSpecName: "utilities") pod "4ae8468f-2d42-4b13-9dbe-f8558df495cc" (UID: "4ae8468f-2d42-4b13-9dbe-f8558df495cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.542456 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ae8468f-2d42-4b13-9dbe-f8558df495cc" (UID: "4ae8468f-2d42-4b13-9dbe-f8558df495cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.543528 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ae8468f-2d42-4b13-9dbe-f8558df495cc-kube-api-access-jdtt2" (OuterVolumeSpecName: "kube-api-access-jdtt2") pod "4ae8468f-2d42-4b13-9dbe-f8558df495cc" (UID: "4ae8468f-2d42-4b13-9dbe-f8558df495cc"). InnerVolumeSpecName "kube-api-access-jdtt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.625806 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.625846 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ae8468f-2d42-4b13-9dbe-f8558df495cc-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.625859 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdtt2\" (UniqueName: \"kubernetes.io/projected/4ae8468f-2d42-4b13-9dbe-f8558df495cc-kube-api-access-jdtt2\") on node \"crc\" DevicePath \"\"" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.721730 4810 generic.go:334] "Generic (PLEG): container finished" podID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" containerID="c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6" exitCode=0 Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.721845 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvbp7" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.721830 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvbp7" event={"ID":"4ae8468f-2d42-4b13-9dbe-f8558df495cc","Type":"ContainerDied","Data":"c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6"} Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.721928 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvbp7" event={"ID":"4ae8468f-2d42-4b13-9dbe-f8558df495cc","Type":"ContainerDied","Data":"f849f484610cac87045cdfbded50bf106ff96e025879b4b97ca1c8d7aa206f09"} Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.721956 4810 scope.go:117] "RemoveContainer" containerID="c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.750740 4810 scope.go:117] "RemoveContainer" containerID="d8c48b26ee828189c3737a97b910929e5a31021340f6793e6a283f3518865358" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.759770 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvbp7"] Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.784545 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvbp7"] Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.799979 4810 scope.go:117] "RemoveContainer" containerID="eb197027db1f95d8ede5d6c85dac7a4ac67c1014b67d6a1ca3bdc9864a3002ec" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.838693 4810 scope.go:117] "RemoveContainer" containerID="c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6" Sep 30 09:49:25 crc kubenswrapper[4810]: E0930 09:49:25.839186 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6\": container with ID starting with c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6 not found: ID does not exist" containerID="c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.839228 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6"} err="failed to get container status \"c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6\": rpc error: code = NotFound desc = could not find container \"c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6\": container with ID starting with c8433311e6fc5fff3ca56564f26f28ba56c31455b19b270a7465a5b30c2c7ef6 not found: ID does not exist" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.839257 4810 scope.go:117] "RemoveContainer" containerID="d8c48b26ee828189c3737a97b910929e5a31021340f6793e6a283f3518865358" Sep 30 09:49:25 crc kubenswrapper[4810]: E0930 09:49:25.839611 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8c48b26ee828189c3737a97b910929e5a31021340f6793e6a283f3518865358\": container with ID starting with d8c48b26ee828189c3737a97b910929e5a31021340f6793e6a283f3518865358 not found: ID does not exist" containerID="d8c48b26ee828189c3737a97b910929e5a31021340f6793e6a283f3518865358" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.839642 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8c48b26ee828189c3737a97b910929e5a31021340f6793e6a283f3518865358"} err="failed to get container status \"d8c48b26ee828189c3737a97b910929e5a31021340f6793e6a283f3518865358\": rpc error: code = NotFound desc = could not find container \"d8c48b26ee828189c3737a97b910929e5a31021340f6793e6a283f3518865358\": container with ID starting with d8c48b26ee828189c3737a97b910929e5a31021340f6793e6a283f3518865358 not found: ID does not exist" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.839660 4810 scope.go:117] "RemoveContainer" containerID="eb197027db1f95d8ede5d6c85dac7a4ac67c1014b67d6a1ca3bdc9864a3002ec" Sep 30 09:49:25 crc kubenswrapper[4810]: E0930 09:49:25.843334 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb197027db1f95d8ede5d6c85dac7a4ac67c1014b67d6a1ca3bdc9864a3002ec\": container with ID starting with eb197027db1f95d8ede5d6c85dac7a4ac67c1014b67d6a1ca3bdc9864a3002ec not found: ID does not exist" containerID="eb197027db1f95d8ede5d6c85dac7a4ac67c1014b67d6a1ca3bdc9864a3002ec" Sep 30 09:49:25 crc kubenswrapper[4810]: I0930 09:49:25.843358 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb197027db1f95d8ede5d6c85dac7a4ac67c1014b67d6a1ca3bdc9864a3002ec"} err="failed to get container status \"eb197027db1f95d8ede5d6c85dac7a4ac67c1014b67d6a1ca3bdc9864a3002ec\": rpc error: code = NotFound desc = could not find container \"eb197027db1f95d8ede5d6c85dac7a4ac67c1014b67d6a1ca3bdc9864a3002ec\": container with ID starting with eb197027db1f95d8ede5d6c85dac7a4ac67c1014b67d6a1ca3bdc9864a3002ec not found: ID does not exist" Sep 30 09:49:27 crc kubenswrapper[4810]: I0930 09:49:27.318701 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" path="/var/lib/kubelet/pods/4ae8468f-2d42-4b13-9dbe-f8558df495cc/volumes" Sep 30 09:49:34 crc kubenswrapper[4810]: I0930 09:49:34.954776 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pdntm/must-gather-2vvzs"] Sep 30 09:49:34 crc kubenswrapper[4810]: I0930 09:49:34.955585 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-pdntm/must-gather-2vvzs" podUID="75f3fd61-26a0-4cb9-bdba-144e953d81e8" containerName="copy" containerID="cri-o://a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560" gracePeriod=2 Sep 30 09:49:34 crc kubenswrapper[4810]: I0930 09:49:34.965031 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pdntm/must-gather-2vvzs"] Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.466800 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pdntm_must-gather-2vvzs_75f3fd61-26a0-4cb9-bdba-144e953d81e8/copy/0.log" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.467536 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/must-gather-2vvzs" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.525893 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mppgh\" (UniqueName: \"kubernetes.io/projected/75f3fd61-26a0-4cb9-bdba-144e953d81e8-kube-api-access-mppgh\") pod \"75f3fd61-26a0-4cb9-bdba-144e953d81e8\" (UID: \"75f3fd61-26a0-4cb9-bdba-144e953d81e8\") " Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.526001 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/75f3fd61-26a0-4cb9-bdba-144e953d81e8-must-gather-output\") pod \"75f3fd61-26a0-4cb9-bdba-144e953d81e8\" (UID: \"75f3fd61-26a0-4cb9-bdba-144e953d81e8\") " Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.531683 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75f3fd61-26a0-4cb9-bdba-144e953d81e8-kube-api-access-mppgh" (OuterVolumeSpecName: "kube-api-access-mppgh") pod "75f3fd61-26a0-4cb9-bdba-144e953d81e8" (UID: "75f3fd61-26a0-4cb9-bdba-144e953d81e8"). InnerVolumeSpecName "kube-api-access-mppgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.627802 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mppgh\" (UniqueName: \"kubernetes.io/projected/75f3fd61-26a0-4cb9-bdba-144e953d81e8-kube-api-access-mppgh\") on node \"crc\" DevicePath \"\"" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.711411 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75f3fd61-26a0-4cb9-bdba-144e953d81e8-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "75f3fd61-26a0-4cb9-bdba-144e953d81e8" (UID: "75f3fd61-26a0-4cb9-bdba-144e953d81e8"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.730432 4810 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/75f3fd61-26a0-4cb9-bdba-144e953d81e8-must-gather-output\") on node \"crc\" DevicePath \"\"" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.808579 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pdntm_must-gather-2vvzs_75f3fd61-26a0-4cb9-bdba-144e953d81e8/copy/0.log" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.810377 4810 generic.go:334] "Generic (PLEG): container finished" podID="75f3fd61-26a0-4cb9-bdba-144e953d81e8" containerID="a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560" exitCode=143 Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.810456 4810 scope.go:117] "RemoveContainer" containerID="a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.810515 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pdntm/must-gather-2vvzs" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.840110 4810 scope.go:117] "RemoveContainer" containerID="fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.902101 4810 scope.go:117] "RemoveContainer" containerID="a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560" Sep 30 09:49:35 crc kubenswrapper[4810]: E0930 09:49:35.902847 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560\": container with ID starting with a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560 not found: ID does not exist" containerID="a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.902914 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560"} err="failed to get container status \"a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560\": rpc error: code = NotFound desc = could not find container \"a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560\": container with ID starting with a2031b61691a0a27894dbce698b925a796a32231206b62d6ccaab57c3c839560 not found: ID does not exist" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.902945 4810 scope.go:117] "RemoveContainer" containerID="fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0" Sep 30 09:49:35 crc kubenswrapper[4810]: E0930 09:49:35.905228 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0\": container with ID starting with fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0 not found: ID does not exist" containerID="fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0" Sep 30 09:49:35 crc kubenswrapper[4810]: I0930 09:49:35.905253 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0"} err="failed to get container status \"fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0\": rpc error: code = NotFound desc = could not find container \"fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0\": container with ID starting with fbf0c72cbb9bc0624c595864eed689158f8cd7ce4d97be8688286cbbdc5935a0 not found: ID does not exist" Sep 30 09:49:37 crc kubenswrapper[4810]: I0930 09:49:37.324900 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75f3fd61-26a0-4cb9-bdba-144e953d81e8" path="/var/lib/kubelet/pods/75f3fd61-26a0-4cb9-bdba-144e953d81e8/volumes" Sep 30 09:50:45 crc kubenswrapper[4810]: I0930 09:50:45.911198 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:50:45 crc kubenswrapper[4810]: I0930 09:50:45.911859 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:51:15 crc kubenswrapper[4810]: I0930 09:51:15.912044 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:51:15 crc kubenswrapper[4810]: I0930 09:51:15.912529 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.429440 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rkbcw"] Sep 30 09:51:16 crc kubenswrapper[4810]: E0930 09:51:16.429830 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" containerName="extract-utilities" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.429844 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" containerName="extract-utilities" Sep 30 09:51:16 crc kubenswrapper[4810]: E0930 09:51:16.429872 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f3fd61-26a0-4cb9-bdba-144e953d81e8" containerName="gather" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.429878 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f3fd61-26a0-4cb9-bdba-144e953d81e8" containerName="gather" Sep 30 09:51:16 crc kubenswrapper[4810]: E0930 09:51:16.429895 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" containerName="extract-content" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.429903 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" containerName="extract-content" Sep 30 09:51:16 crc kubenswrapper[4810]: E0930 09:51:16.429916 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" containerName="registry-server" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.429922 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" containerName="registry-server" Sep 30 09:51:16 crc kubenswrapper[4810]: E0930 09:51:16.429942 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f3fd61-26a0-4cb9-bdba-144e953d81e8" containerName="copy" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.429948 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f3fd61-26a0-4cb9-bdba-144e953d81e8" containerName="copy" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.430139 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="75f3fd61-26a0-4cb9-bdba-144e953d81e8" containerName="copy" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.430186 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ae8468f-2d42-4b13-9dbe-f8558df495cc" containerName="registry-server" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.430207 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="75f3fd61-26a0-4cb9-bdba-144e953d81e8" containerName="gather" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.432329 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.441882 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rkbcw"] Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.540418 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm8dz\" (UniqueName: \"kubernetes.io/projected/aa43e924-98f3-4cea-b0fa-870659d14325-kube-api-access-rm8dz\") pod \"certified-operators-rkbcw\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.540500 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-utilities\") pod \"certified-operators-rkbcw\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.540543 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-catalog-content\") pod \"certified-operators-rkbcw\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.642144 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm8dz\" (UniqueName: \"kubernetes.io/projected/aa43e924-98f3-4cea-b0fa-870659d14325-kube-api-access-rm8dz\") pod \"certified-operators-rkbcw\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.642213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-utilities\") pod \"certified-operators-rkbcw\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.642240 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-catalog-content\") pod \"certified-operators-rkbcw\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.642869 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-utilities\") pod \"certified-operators-rkbcw\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.642960 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-catalog-content\") pod \"certified-operators-rkbcw\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.669705 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm8dz\" (UniqueName: \"kubernetes.io/projected/aa43e924-98f3-4cea-b0fa-870659d14325-kube-api-access-rm8dz\") pod \"certified-operators-rkbcw\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:16 crc kubenswrapper[4810]: I0930 09:51:16.757407 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:17 crc kubenswrapper[4810]: I0930 09:51:17.335311 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rkbcw"] Sep 30 09:51:17 crc kubenswrapper[4810]: I0930 09:51:17.851521 4810 generic.go:334] "Generic (PLEG): container finished" podID="aa43e924-98f3-4cea-b0fa-870659d14325" containerID="6e427a4238fd704b037a36ae669fdb8125d5ec80f4a203acbdba46d8333e1b4a" exitCode=0 Sep 30 09:51:17 crc kubenswrapper[4810]: I0930 09:51:17.851589 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkbcw" event={"ID":"aa43e924-98f3-4cea-b0fa-870659d14325","Type":"ContainerDied","Data":"6e427a4238fd704b037a36ae669fdb8125d5ec80f4a203acbdba46d8333e1b4a"} Sep 30 09:51:17 crc kubenswrapper[4810]: I0930 09:51:17.851817 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkbcw" event={"ID":"aa43e924-98f3-4cea-b0fa-870659d14325","Type":"ContainerStarted","Data":"04fbec38b224006188d227a8648e9752f77646b1e8f9d87a8169b2360e6da311"} Sep 30 09:51:19 crc kubenswrapper[4810]: I0930 09:51:19.874878 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkbcw" event={"ID":"aa43e924-98f3-4cea-b0fa-870659d14325","Type":"ContainerStarted","Data":"14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222"} Sep 30 09:51:20 crc kubenswrapper[4810]: I0930 09:51:20.886672 4810 generic.go:334] "Generic (PLEG): container finished" podID="aa43e924-98f3-4cea-b0fa-870659d14325" containerID="14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222" exitCode=0 Sep 30 09:51:20 crc kubenswrapper[4810]: I0930 09:51:20.886725 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkbcw" event={"ID":"aa43e924-98f3-4cea-b0fa-870659d14325","Type":"ContainerDied","Data":"14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222"} Sep 30 09:51:21 crc kubenswrapper[4810]: I0930 09:51:21.898843 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkbcw" event={"ID":"aa43e924-98f3-4cea-b0fa-870659d14325","Type":"ContainerStarted","Data":"297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905"} Sep 30 09:51:21 crc kubenswrapper[4810]: I0930 09:51:21.930855 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rkbcw" podStartSLOduration=2.453065875 podStartE2EDuration="5.930833598s" podCreationTimestamp="2025-09-30 09:51:16 +0000 UTC" firstStartedPulling="2025-09-30 09:51:17.854534102 +0000 UTC m=+6501.306733369" lastFinishedPulling="2025-09-30 09:51:21.332301835 +0000 UTC m=+6504.784501092" observedRunningTime="2025-09-30 09:51:21.919292954 +0000 UTC m=+6505.371492231" watchObservedRunningTime="2025-09-30 09:51:21.930833598 +0000 UTC m=+6505.383032865" Sep 30 09:51:24 crc kubenswrapper[4810]: I0930 09:51:24.065068 4810 scope.go:117] "RemoveContainer" containerID="dedcbbb66df6bf7b2242960a1e62a57499b8ed93542c8b1041e885fa4651c6ee" Sep 30 09:51:26 crc kubenswrapper[4810]: I0930 09:51:26.758154 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:26 crc kubenswrapper[4810]: I0930 09:51:26.758491 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:26 crc kubenswrapper[4810]: I0930 09:51:26.807025 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:26 crc kubenswrapper[4810]: I0930 09:51:26.981583 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:27 crc kubenswrapper[4810]: I0930 09:51:27.039025 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rkbcw"] Sep 30 09:51:28 crc kubenswrapper[4810]: I0930 09:51:28.954966 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rkbcw" podUID="aa43e924-98f3-4cea-b0fa-870659d14325" containerName="registry-server" containerID="cri-o://297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905" gracePeriod=2 Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.380628 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.443967 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-utilities\") pod \"aa43e924-98f3-4cea-b0fa-870659d14325\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.444097 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-catalog-content\") pod \"aa43e924-98f3-4cea-b0fa-870659d14325\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.444167 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm8dz\" (UniqueName: \"kubernetes.io/projected/aa43e924-98f3-4cea-b0fa-870659d14325-kube-api-access-rm8dz\") pod \"aa43e924-98f3-4cea-b0fa-870659d14325\" (UID: \"aa43e924-98f3-4cea-b0fa-870659d14325\") " Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.445185 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-utilities" (OuterVolumeSpecName: "utilities") pod "aa43e924-98f3-4cea-b0fa-870659d14325" (UID: "aa43e924-98f3-4cea-b0fa-870659d14325"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.453682 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa43e924-98f3-4cea-b0fa-870659d14325-kube-api-access-rm8dz" (OuterVolumeSpecName: "kube-api-access-rm8dz") pod "aa43e924-98f3-4cea-b0fa-870659d14325" (UID: "aa43e924-98f3-4cea-b0fa-870659d14325"). InnerVolumeSpecName "kube-api-access-rm8dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.499788 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa43e924-98f3-4cea-b0fa-870659d14325" (UID: "aa43e924-98f3-4cea-b0fa-870659d14325"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.546631 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.546691 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm8dz\" (UniqueName: \"kubernetes.io/projected/aa43e924-98f3-4cea-b0fa-870659d14325-kube-api-access-rm8dz\") on node \"crc\" DevicePath \"\"" Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.546706 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa43e924-98f3-4cea-b0fa-870659d14325-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.969203 4810 generic.go:334] "Generic (PLEG): container finished" podID="aa43e924-98f3-4cea-b0fa-870659d14325" containerID="297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905" exitCode=0 Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.969327 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkbcw" event={"ID":"aa43e924-98f3-4cea-b0fa-870659d14325","Type":"ContainerDied","Data":"297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905"} Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.969880 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkbcw" event={"ID":"aa43e924-98f3-4cea-b0fa-870659d14325","Type":"ContainerDied","Data":"04fbec38b224006188d227a8648e9752f77646b1e8f9d87a8169b2360e6da311"} Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.969927 4810 scope.go:117] "RemoveContainer" containerID="297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905" Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.970415 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkbcw" Sep 30 09:51:29 crc kubenswrapper[4810]: I0930 09:51:29.991524 4810 scope.go:117] "RemoveContainer" containerID="14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222" Sep 30 09:51:30 crc kubenswrapper[4810]: I0930 09:51:30.001919 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rkbcw"] Sep 30 09:51:30 crc kubenswrapper[4810]: I0930 09:51:30.010068 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rkbcw"] Sep 30 09:51:30 crc kubenswrapper[4810]: I0930 09:51:30.026250 4810 scope.go:117] "RemoveContainer" containerID="6e427a4238fd704b037a36ae669fdb8125d5ec80f4a203acbdba46d8333e1b4a" Sep 30 09:51:30 crc kubenswrapper[4810]: I0930 09:51:30.062828 4810 scope.go:117] "RemoveContainer" containerID="297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905" Sep 30 09:51:30 crc kubenswrapper[4810]: E0930 09:51:30.063137 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905\": container with ID starting with 297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905 not found: ID does not exist" containerID="297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905" Sep 30 09:51:30 crc kubenswrapper[4810]: I0930 09:51:30.063168 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905"} err="failed to get container status \"297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905\": rpc error: code = NotFound desc = could not find container \"297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905\": container with ID starting with 297665199d52a20b6d28bbad24928039f84f6ce4e0a1b50fd32153b63acfd905 not found: ID does not exist" Sep 30 09:51:30 crc kubenswrapper[4810]: I0930 09:51:30.063186 4810 scope.go:117] "RemoveContainer" containerID="14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222" Sep 30 09:51:30 crc kubenswrapper[4810]: E0930 09:51:30.063420 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222\": container with ID starting with 14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222 not found: ID does not exist" containerID="14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222" Sep 30 09:51:30 crc kubenswrapper[4810]: I0930 09:51:30.063445 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222"} err="failed to get container status \"14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222\": rpc error: code = NotFound desc = could not find container \"14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222\": container with ID starting with 14d6cd861e363a13b73b419333ccac7d1cecd6cac90b8f3d93ea001597c27222 not found: ID does not exist" Sep 30 09:51:30 crc kubenswrapper[4810]: I0930 09:51:30.063461 4810 scope.go:117] "RemoveContainer" containerID="6e427a4238fd704b037a36ae669fdb8125d5ec80f4a203acbdba46d8333e1b4a" Sep 30 09:51:30 crc kubenswrapper[4810]: E0930 09:51:30.063745 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e427a4238fd704b037a36ae669fdb8125d5ec80f4a203acbdba46d8333e1b4a\": container with ID starting with 6e427a4238fd704b037a36ae669fdb8125d5ec80f4a203acbdba46d8333e1b4a not found: ID does not exist" containerID="6e427a4238fd704b037a36ae669fdb8125d5ec80f4a203acbdba46d8333e1b4a" Sep 30 09:51:30 crc kubenswrapper[4810]: I0930 09:51:30.063778 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e427a4238fd704b037a36ae669fdb8125d5ec80f4a203acbdba46d8333e1b4a"} err="failed to get container status \"6e427a4238fd704b037a36ae669fdb8125d5ec80f4a203acbdba46d8333e1b4a\": rpc error: code = NotFound desc = could not find container \"6e427a4238fd704b037a36ae669fdb8125d5ec80f4a203acbdba46d8333e1b4a\": container with ID starting with 6e427a4238fd704b037a36ae669fdb8125d5ec80f4a203acbdba46d8333e1b4a not found: ID does not exist" Sep 30 09:51:31 crc kubenswrapper[4810]: I0930 09:51:31.319404 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa43e924-98f3-4cea-b0fa-870659d14325" path="/var/lib/kubelet/pods/aa43e924-98f3-4cea-b0fa-870659d14325/volumes" Sep 30 09:51:45 crc kubenswrapper[4810]: I0930 09:51:45.911661 4810 patch_prober.go:28] interesting pod/machine-config-daemon-hk2k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 09:51:45 crc kubenswrapper[4810]: I0930 09:51:45.912186 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 09:51:45 crc kubenswrapper[4810]: I0930 09:51:45.912239 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" Sep 30 09:51:45 crc kubenswrapper[4810]: I0930 09:51:45.913050 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"19649ce9d9c97dcaa82fd170c48b84ab4529eb7ee9a0d6d010bcfc9f2397e8d8"} pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 09:51:45 crc kubenswrapper[4810]: I0930 09:51:45.913109 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" podUID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerName="machine-config-daemon" containerID="cri-o://19649ce9d9c97dcaa82fd170c48b84ab4529eb7ee9a0d6d010bcfc9f2397e8d8" gracePeriod=600 Sep 30 09:51:46 crc kubenswrapper[4810]: I0930 09:51:46.137084 4810 generic.go:334] "Generic (PLEG): container finished" podID="86194959-30fe-4da8-a64b-f5a849c2f73a" containerID="19649ce9d9c97dcaa82fd170c48b84ab4529eb7ee9a0d6d010bcfc9f2397e8d8" exitCode=0 Sep 30 09:51:46 crc kubenswrapper[4810]: I0930 09:51:46.137172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerDied","Data":"19649ce9d9c97dcaa82fd170c48b84ab4529eb7ee9a0d6d010bcfc9f2397e8d8"} Sep 30 09:51:46 crc kubenswrapper[4810]: I0930 09:51:46.137396 4810 scope.go:117] "RemoveContainer" containerID="7f4e4e4dc806a1d4aae0ce6ce0963bf806286ad31625146cb5b7c7b7ad78409e" Sep 30 09:51:47 crc kubenswrapper[4810]: I0930 09:51:47.149066 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hk2k2" event={"ID":"86194959-30fe-4da8-a64b-f5a849c2f73a","Type":"ContainerStarted","Data":"a7039e4bc492518db012e2235806f3efd570c52ea502634791ae78be0f109151"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515066724077024463 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015066724077017400 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015066706672016524 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015066706672015474 5ustar corecore